perf_counter, x86: Add mmap counter read support
authorPeter Zijlstra <a.p.zijlstra@chello.nl>
Mon, 22 Jun 2009 14:35:24 +0000 (16:35 +0200)
committerIngo Molnar <mingo@elte.hu>
Thu, 25 Jun 2009 19:39:06 +0000 (21:39 +0200)
Update the mmap control page with the needed information to
use the userspace RDPMC instruction for self monitoring.

Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
LKML-Reference: <new-submission>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
arch/powerpc/include/asm/perf_counter.h
arch/x86/include/asm/perf_counter.h
arch/x86/kernel/cpu/perf_counter.c
kernel/perf_counter.c

index 8ccd4e155768fa749edf3f65fbc0c5a951ef9819..0ea0639fcf75be1c4393e4437aceaf7bb7ca87eb 100644 (file)
@@ -61,6 +61,8 @@ struct pt_regs;
 extern unsigned long perf_misc_flags(struct pt_regs *regs);
 extern unsigned long perf_instruction_pointer(struct pt_regs *regs);
 
+#define PERF_COUNTER_INDEX_OFFSET      1
+
 /*
  * Only override the default definitions in include/linux/perf_counter.h
  * if we have hardware PMU support.
index 5fb33e160ea0b585b065b726ede5a855bfa22df7..fa64e401589d4ed9089c908f06c3bef2261c4d9f 100644 (file)
@@ -87,6 +87,9 @@ union cpuid10_edx {
 #ifdef CONFIG_PERF_COUNTERS
 extern void init_hw_perf_counters(void);
 extern void perf_counters_lapic_init(void);
+
+#define PERF_COUNTER_INDEX_OFFSET                      0
+
 #else
 static inline void init_hw_perf_counters(void)         { }
 static inline void perf_counters_lapic_init(void)      { }
index a310d19faca36ae8ef0fd1f25ad92145ae7f3ea9..b83474b6021ad6a93b8747b861d29dd3691c9d92 100644 (file)
@@ -912,6 +912,8 @@ x86_perf_counter_set_period(struct perf_counter *counter,
        err = checking_wrmsrl(hwc->counter_base + idx,
                             (u64)(-left) & x86_pmu.counter_mask);
 
+       perf_counter_update_userpage(counter);
+
        return ret;
 }
 
@@ -1034,6 +1036,8 @@ try_generic:
        x86_perf_counter_set_period(counter, hwc, idx);
        x86_pmu.enable(hwc, idx);
 
+       perf_counter_update_userpage(counter);
+
        return 0;
 }
 
@@ -1126,6 +1130,8 @@ static void x86_pmu_disable(struct perf_counter *counter)
        x86_perf_counter_update(counter, hwc, idx);
        cpuc->counters[idx] = NULL;
        clear_bit(idx, cpuc->used_mask);
+
+       perf_counter_update_userpage(counter);
 }
 
 /*
index 23614adab475b6fa7c4dfa90fb78dc1b7f9cd843..02994a719e2711f4a09de6b96f02d87ec92899af 100644 (file)
@@ -1753,6 +1753,14 @@ int perf_counter_task_disable(void)
        return 0;
 }
 
+static int perf_counter_index(struct perf_counter *counter)
+{
+       if (counter->state != PERF_COUNTER_STATE_ACTIVE)
+               return 0;
+
+       return counter->hw.idx + 1 - PERF_COUNTER_INDEX_OFFSET;
+}
+
 /*
  * Callers need to ensure there can be no nesting of this function, otherwise
  * the seqlock logic goes bad. We can not serialize this because the arch
@@ -1777,7 +1785,7 @@ void perf_counter_update_userpage(struct perf_counter *counter)
        preempt_disable();
        ++userpg->lock;
        barrier();
-       userpg->index = counter->hw.idx;
+       userpg->index = perf_counter_index(counter);
        userpg->offset = atomic64_read(&counter->count);
        if (counter->state == PERF_COUNTER_STATE_ACTIVE)
                userpg->offset -= atomic64_read(&counter->hw.prev_count);