forked from Minki/linux
90eec103b9
There were still a number of references to my old Red Hat email address in the kernel source. Remove these while keeping the Red Hat copyright notices intact. Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Cc: Arnaldo Carvalho de Melo <acme@redhat.com> Cc: Jiri Olsa <jolsa@redhat.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Mike Galbraith <efault@gmx.de> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Stephane Eranian <eranian@google.com> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Vince Weaver <vincent.weaver@maine.edu> Signed-off-by: Ingo Molnar <mingo@kernel.org>
1006 lines
24 KiB
C
1006 lines
24 KiB
C
/*
|
|
* Copyright 2014 Tilera Corporation. All Rights Reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation, version 2.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
|
|
* NON INFRINGEMENT. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
*
|
|
* Perf_events support for Tile processor.
|
|
*
|
|
* This code is based upon the x86 perf event
|
|
* code, which is:
|
|
*
|
|
* Copyright (C) 2008 Thomas Gleixner <tglx@linutronix.de>
|
|
* Copyright (C) 2008-2009 Red Hat, Inc., Ingo Molnar
|
|
* Copyright (C) 2009 Jaswinder Singh Rajput
|
|
* Copyright (C) 2009 Advanced Micro Devices, Inc., Robert Richter
|
|
* Copyright (C) 2008-2009 Red Hat, Inc., Peter Zijlstra
|
|
* Copyright (C) 2009 Intel Corporation, <markus.t.metzger@intel.com>
|
|
* Copyright (C) 2009 Google, Inc., Stephane Eranian
|
|
*/
|
|
|
|
#include <linux/kprobes.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/kdebug.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/bitmap.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/perf_event.h>
|
|
#include <linux/atomic.h>
|
|
#include <asm/traps.h>
|
|
#include <asm/stack.h>
|
|
#include <asm/pmc.h>
|
|
#include <hv/hypervisor.h>
|
|
|
|
#define TILE_MAX_COUNTERS 4
|
|
|
|
#define PERF_COUNT_0_IDX 0
|
|
#define PERF_COUNT_1_IDX 1
|
|
#define AUX_PERF_COUNT_0_IDX 2
|
|
#define AUX_PERF_COUNT_1_IDX 3
|
|
|
|
struct cpu_hw_events {
|
|
int n_events;
|
|
struct perf_event *events[TILE_MAX_COUNTERS]; /* counter order */
|
|
struct perf_event *event_list[TILE_MAX_COUNTERS]; /* enabled
|
|
order */
|
|
int assign[TILE_MAX_COUNTERS];
|
|
unsigned long active_mask[BITS_TO_LONGS(TILE_MAX_COUNTERS)];
|
|
unsigned long used_mask;
|
|
};
|
|
|
|
/* TILE arch specific performance monitor unit */
|
|
struct tile_pmu {
|
|
const char *name;
|
|
int version;
|
|
const int *hw_events; /* generic hw events table */
|
|
/* generic hw cache events table */
|
|
const int (*cache_events)[PERF_COUNT_HW_CACHE_MAX]
|
|
[PERF_COUNT_HW_CACHE_OP_MAX]
|
|
[PERF_COUNT_HW_CACHE_RESULT_MAX];
|
|
int (*map_hw_event)(u64); /*method used to map
|
|
hw events */
|
|
int (*map_cache_event)(u64); /*method used to map
|
|
cache events */
|
|
|
|
u64 max_period; /* max sampling period */
|
|
u64 cntval_mask; /* counter width mask */
|
|
int cntval_bits; /* counter width */
|
|
int max_events; /* max generic hw events
|
|
in map */
|
|
int num_counters; /* number base + aux counters */
|
|
int num_base_counters; /* number base counters */
|
|
};
|
|
|
|
DEFINE_PER_CPU(u64, perf_irqs);
|
|
static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events);
|
|
|
|
#define TILE_OP_UNSUPP (-1)
|
|
|
|
#ifndef __tilegx__
|
|
/* TILEPro hardware events map */
|
|
static const int tile_hw_event_map[] = {
|
|
[PERF_COUNT_HW_CPU_CYCLES] = 0x01, /* ONE */
|
|
[PERF_COUNT_HW_INSTRUCTIONS] = 0x06, /* MP_BUNDLE_RETIRED */
|
|
[PERF_COUNT_HW_CACHE_REFERENCES] = TILE_OP_UNSUPP,
|
|
[PERF_COUNT_HW_CACHE_MISSES] = TILE_OP_UNSUPP,
|
|
[PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = 0x16, /*
|
|
MP_CONDITIONAL_BRANCH_ISSUED */
|
|
[PERF_COUNT_HW_BRANCH_MISSES] = 0x14, /*
|
|
MP_CONDITIONAL_BRANCH_MISSPREDICT */
|
|
[PERF_COUNT_HW_BUS_CYCLES] = TILE_OP_UNSUPP,
|
|
};
|
|
#else
|
|
/* TILEGx hardware events map */
|
|
static const int tile_hw_event_map[] = {
|
|
[PERF_COUNT_HW_CPU_CYCLES] = 0x181, /* ONE */
|
|
[PERF_COUNT_HW_INSTRUCTIONS] = 0xdb, /* INSTRUCTION_BUNDLE */
|
|
[PERF_COUNT_HW_CACHE_REFERENCES] = TILE_OP_UNSUPP,
|
|
[PERF_COUNT_HW_CACHE_MISSES] = TILE_OP_UNSUPP,
|
|
[PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = 0xd9, /*
|
|
COND_BRANCH_PRED_CORRECT */
|
|
[PERF_COUNT_HW_BRANCH_MISSES] = 0xda, /*
|
|
COND_BRANCH_PRED_INCORRECT */
|
|
[PERF_COUNT_HW_BUS_CYCLES] = TILE_OP_UNSUPP,
|
|
};
|
|
#endif
|
|
|
|
#define C(x) PERF_COUNT_HW_CACHE_##x
|
|
|
|
/*
|
|
* Generalized hw caching related hw_event table, filled
|
|
* in on a per model basis. A value of -1 means
|
|
* 'not supported', any other value means the
|
|
* raw hw_event ID.
|
|
*/
|
|
#ifndef __tilegx__
|
|
/* TILEPro hardware cache event map */
|
|
static const int tile_cache_event_map[PERF_COUNT_HW_CACHE_MAX]
|
|
[PERF_COUNT_HW_CACHE_OP_MAX]
|
|
[PERF_COUNT_HW_CACHE_RESULT_MAX] = {
|
|
[C(L1D)] = {
|
|
[C(OP_READ)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = 0x21, /* RD_MISS */
|
|
},
|
|
[C(OP_WRITE)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = 0x22, /* WR_MISS */
|
|
},
|
|
[C(OP_PREFETCH)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
},
|
|
[C(L1I)] = {
|
|
[C(OP_READ)] = {
|
|
[C(RESULT_ACCESS)] = 0x12, /* MP_ICACHE_HIT_ISSUED */
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_WRITE)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_PREFETCH)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
},
|
|
[C(LL)] = {
|
|
[C(OP_READ)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_WRITE)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_PREFETCH)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
},
|
|
[C(DTLB)] = {
|
|
[C(OP_READ)] = {
|
|
[C(RESULT_ACCESS)] = 0x1d, /* TLB_CNT */
|
|
[C(RESULT_MISS)] = 0x20, /* TLB_EXCEPTION */
|
|
},
|
|
[C(OP_WRITE)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_PREFETCH)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
},
|
|
[C(ITLB)] = {
|
|
[C(OP_READ)] = {
|
|
[C(RESULT_ACCESS)] = 0x13, /* MP_ITLB_HIT_ISSUED */
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_WRITE)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_PREFETCH)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
},
|
|
[C(BPU)] = {
|
|
[C(OP_READ)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_WRITE)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_PREFETCH)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
},
|
|
};
|
|
#else
|
|
/* TILEGx hardware events map */
|
|
static const int tile_cache_event_map[PERF_COUNT_HW_CACHE_MAX]
|
|
[PERF_COUNT_HW_CACHE_OP_MAX]
|
|
[PERF_COUNT_HW_CACHE_RESULT_MAX] = {
|
|
[C(L1D)] = {
|
|
/*
|
|
* Like some other architectures (e.g. ARM), the performance
|
|
* counters don't differentiate between read and write
|
|
* accesses/misses, so this isn't strictly correct, but it's the
|
|
* best we can do. Writes and reads get combined.
|
|
*/
|
|
[C(OP_READ)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = 0x44, /* RD_MISS */
|
|
},
|
|
[C(OP_WRITE)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = 0x45, /* WR_MISS */
|
|
},
|
|
[C(OP_PREFETCH)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
},
|
|
[C(L1I)] = {
|
|
[C(OP_READ)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_WRITE)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_PREFETCH)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
},
|
|
[C(LL)] = {
|
|
[C(OP_READ)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_WRITE)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_PREFETCH)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
},
|
|
[C(DTLB)] = {
|
|
[C(OP_READ)] = {
|
|
[C(RESULT_ACCESS)] = 0x40, /* TLB_CNT */
|
|
[C(RESULT_MISS)] = 0x43, /* TLB_EXCEPTION */
|
|
},
|
|
[C(OP_WRITE)] = {
|
|
[C(RESULT_ACCESS)] = 0x40, /* TLB_CNT */
|
|
[C(RESULT_MISS)] = 0x43, /* TLB_EXCEPTION */
|
|
},
|
|
[C(OP_PREFETCH)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
},
|
|
[C(ITLB)] = {
|
|
[C(OP_READ)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = 0xd4, /* ITLB_MISS_INT */
|
|
},
|
|
[C(OP_WRITE)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = 0xd4, /* ITLB_MISS_INT */
|
|
},
|
|
[C(OP_PREFETCH)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
},
|
|
[C(BPU)] = {
|
|
[C(OP_READ)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_WRITE)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
[C(OP_PREFETCH)] = {
|
|
[C(RESULT_ACCESS)] = TILE_OP_UNSUPP,
|
|
[C(RESULT_MISS)] = TILE_OP_UNSUPP,
|
|
},
|
|
},
|
|
};
|
|
#endif
|
|
|
|
static atomic_t tile_active_events;
|
|
static DEFINE_MUTEX(perf_intr_reserve_mutex);
|
|
|
|
static int tile_map_hw_event(u64 config);
|
|
static int tile_map_cache_event(u64 config);
|
|
|
|
static int tile_pmu_handle_irq(struct pt_regs *regs, int fault);
|
|
|
|
/*
|
|
* To avoid new_raw_count getting larger then pre_raw_count
|
|
* in tile_perf_event_update(), we limit the value of max_period to 2^31 - 1.
|
|
*/
|
|
static const struct tile_pmu tilepmu = {
|
|
#ifndef __tilegx__
|
|
.name = "tilepro",
|
|
#else
|
|
.name = "tilegx",
|
|
#endif
|
|
.max_events = ARRAY_SIZE(tile_hw_event_map),
|
|
.map_hw_event = tile_map_hw_event,
|
|
.hw_events = tile_hw_event_map,
|
|
.map_cache_event = tile_map_cache_event,
|
|
.cache_events = &tile_cache_event_map,
|
|
.cntval_bits = 32,
|
|
.cntval_mask = (1ULL << 32) - 1,
|
|
.max_period = (1ULL << 31) - 1,
|
|
.num_counters = TILE_MAX_COUNTERS,
|
|
.num_base_counters = TILE_BASE_COUNTERS,
|
|
};
|
|
|
|
static const struct tile_pmu *tile_pmu __read_mostly;
|
|
|
|
/*
|
|
* Check whether perf event is enabled.
|
|
*/
|
|
int tile_perf_enabled(void)
|
|
{
|
|
return atomic_read(&tile_active_events) != 0;
|
|
}
|
|
|
|
/*
|
|
* Read Performance Counters.
|
|
*/
|
|
static inline u64 read_counter(int idx)
|
|
{
|
|
u64 val = 0;
|
|
|
|
/* __insn_mfspr() only takes an immediate argument */
|
|
switch (idx) {
|
|
case PERF_COUNT_0_IDX:
|
|
val = __insn_mfspr(SPR_PERF_COUNT_0);
|
|
break;
|
|
case PERF_COUNT_1_IDX:
|
|
val = __insn_mfspr(SPR_PERF_COUNT_1);
|
|
break;
|
|
case AUX_PERF_COUNT_0_IDX:
|
|
val = __insn_mfspr(SPR_AUX_PERF_COUNT_0);
|
|
break;
|
|
case AUX_PERF_COUNT_1_IDX:
|
|
val = __insn_mfspr(SPR_AUX_PERF_COUNT_1);
|
|
break;
|
|
default:
|
|
WARN_ON_ONCE(idx > AUX_PERF_COUNT_1_IDX ||
|
|
idx < PERF_COUNT_0_IDX);
|
|
}
|
|
|
|
return val;
|
|
}
|
|
|
|
/*
|
|
* Write Performance Counters.
|
|
*/
|
|
static inline void write_counter(int idx, u64 value)
|
|
{
|
|
/* __insn_mtspr() only takes an immediate argument */
|
|
switch (idx) {
|
|
case PERF_COUNT_0_IDX:
|
|
__insn_mtspr(SPR_PERF_COUNT_0, value);
|
|
break;
|
|
case PERF_COUNT_1_IDX:
|
|
__insn_mtspr(SPR_PERF_COUNT_1, value);
|
|
break;
|
|
case AUX_PERF_COUNT_0_IDX:
|
|
__insn_mtspr(SPR_AUX_PERF_COUNT_0, value);
|
|
break;
|
|
case AUX_PERF_COUNT_1_IDX:
|
|
__insn_mtspr(SPR_AUX_PERF_COUNT_1, value);
|
|
break;
|
|
default:
|
|
WARN_ON_ONCE(idx > AUX_PERF_COUNT_1_IDX ||
|
|
idx < PERF_COUNT_0_IDX);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Enable performance event by setting
|
|
* Performance Counter Control registers.
|
|
*/
|
|
static inline void tile_pmu_enable_event(struct perf_event *event)
|
|
{
|
|
struct hw_perf_event *hwc = &event->hw;
|
|
unsigned long cfg, mask;
|
|
int shift, idx = hwc->idx;
|
|
|
|
/*
|
|
* prevent early activation from tile_pmu_start() in hw_perf_enable
|
|
*/
|
|
|
|
if (WARN_ON_ONCE(idx == -1))
|
|
return;
|
|
|
|
if (idx < tile_pmu->num_base_counters)
|
|
cfg = __insn_mfspr(SPR_PERF_COUNT_CTL);
|
|
else
|
|
cfg = __insn_mfspr(SPR_AUX_PERF_COUNT_CTL);
|
|
|
|
switch (idx) {
|
|
case PERF_COUNT_0_IDX:
|
|
case AUX_PERF_COUNT_0_IDX:
|
|
mask = TILE_EVENT_MASK;
|
|
shift = 0;
|
|
break;
|
|
case PERF_COUNT_1_IDX:
|
|
case AUX_PERF_COUNT_1_IDX:
|
|
mask = TILE_EVENT_MASK << 16;
|
|
shift = 16;
|
|
break;
|
|
default:
|
|
WARN_ON_ONCE(idx < PERF_COUNT_0_IDX ||
|
|
idx > AUX_PERF_COUNT_1_IDX);
|
|
return;
|
|
}
|
|
|
|
/* Clear mask bits to enable the event. */
|
|
cfg &= ~mask;
|
|
cfg |= hwc->config << shift;
|
|
|
|
if (idx < tile_pmu->num_base_counters)
|
|
__insn_mtspr(SPR_PERF_COUNT_CTL, cfg);
|
|
else
|
|
__insn_mtspr(SPR_AUX_PERF_COUNT_CTL, cfg);
|
|
}
|
|
|
|
/*
|
|
* Disable performance event by clearing
|
|
* Performance Counter Control registers.
|
|
*/
|
|
static inline void tile_pmu_disable_event(struct perf_event *event)
|
|
{
|
|
struct hw_perf_event *hwc = &event->hw;
|
|
unsigned long cfg, mask;
|
|
int idx = hwc->idx;
|
|
|
|
if (idx == -1)
|
|
return;
|
|
|
|
if (idx < tile_pmu->num_base_counters)
|
|
cfg = __insn_mfspr(SPR_PERF_COUNT_CTL);
|
|
else
|
|
cfg = __insn_mfspr(SPR_AUX_PERF_COUNT_CTL);
|
|
|
|
switch (idx) {
|
|
case PERF_COUNT_0_IDX:
|
|
case AUX_PERF_COUNT_0_IDX:
|
|
mask = TILE_PLM_MASK;
|
|
break;
|
|
case PERF_COUNT_1_IDX:
|
|
case AUX_PERF_COUNT_1_IDX:
|
|
mask = TILE_PLM_MASK << 16;
|
|
break;
|
|
default:
|
|
WARN_ON_ONCE(idx < PERF_COUNT_0_IDX ||
|
|
idx > AUX_PERF_COUNT_1_IDX);
|
|
return;
|
|
}
|
|
|
|
/* Set mask bits to disable the event. */
|
|
cfg |= mask;
|
|
|
|
if (idx < tile_pmu->num_base_counters)
|
|
__insn_mtspr(SPR_PERF_COUNT_CTL, cfg);
|
|
else
|
|
__insn_mtspr(SPR_AUX_PERF_COUNT_CTL, cfg);
|
|
}
|
|
|
|
/*
|
|
* Propagate event elapsed time into the generic event.
|
|
* Can only be executed on the CPU where the event is active.
|
|
* Returns the delta events processed.
|
|
*/
|
|
static u64 tile_perf_event_update(struct perf_event *event)
|
|
{
|
|
struct hw_perf_event *hwc = &event->hw;
|
|
int shift = 64 - tile_pmu->cntval_bits;
|
|
u64 prev_raw_count, new_raw_count;
|
|
u64 oldval;
|
|
int idx = hwc->idx;
|
|
u64 delta;
|
|
|
|
/*
|
|
* Careful: an NMI might modify the previous event value.
|
|
*
|
|
* Our tactic to handle this is to first atomically read and
|
|
* exchange a new raw count - then add that new-prev delta
|
|
* count to the generic event atomically:
|
|
*/
|
|
again:
|
|
prev_raw_count = local64_read(&hwc->prev_count);
|
|
new_raw_count = read_counter(idx);
|
|
|
|
oldval = local64_cmpxchg(&hwc->prev_count, prev_raw_count,
|
|
new_raw_count);
|
|
if (oldval != prev_raw_count)
|
|
goto again;
|
|
|
|
/*
|
|
* Now we have the new raw value and have updated the prev
|
|
* timestamp already. We can now calculate the elapsed delta
|
|
* (event-)time and add that to the generic event.
|
|
*
|
|
* Careful, not all hw sign-extends above the physical width
|
|
* of the count.
|
|
*/
|
|
delta = (new_raw_count << shift) - (prev_raw_count << shift);
|
|
delta >>= shift;
|
|
|
|
local64_add(delta, &event->count);
|
|
local64_sub(delta, &hwc->period_left);
|
|
|
|
return new_raw_count;
|
|
}
|
|
|
|
/*
|
|
* Set the next IRQ period, based on the hwc->period_left value.
|
|
* To be called with the event disabled in hw:
|
|
*/
|
|
static int tile_event_set_period(struct perf_event *event)
|
|
{
|
|
struct hw_perf_event *hwc = &event->hw;
|
|
int idx = hwc->idx;
|
|
s64 left = local64_read(&hwc->period_left);
|
|
s64 period = hwc->sample_period;
|
|
int ret = 0;
|
|
|
|
/*
|
|
* If we are way outside a reasonable range then just skip forward:
|
|
*/
|
|
if (unlikely(left <= -period)) {
|
|
left = period;
|
|
local64_set(&hwc->period_left, left);
|
|
hwc->last_period = period;
|
|
ret = 1;
|
|
}
|
|
|
|
if (unlikely(left <= 0)) {
|
|
left += period;
|
|
local64_set(&hwc->period_left, left);
|
|
hwc->last_period = period;
|
|
ret = 1;
|
|
}
|
|
if (left > tile_pmu->max_period)
|
|
left = tile_pmu->max_period;
|
|
|
|
/*
|
|
* The hw event starts counting from this event offset,
|
|
* mark it to be able to extra future deltas:
|
|
*/
|
|
local64_set(&hwc->prev_count, (u64)-left);
|
|
|
|
write_counter(idx, (u64)(-left) & tile_pmu->cntval_mask);
|
|
|
|
perf_event_update_userpage(event);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Stop the event but do not release the PMU counter
|
|
*/
|
|
static void tile_pmu_stop(struct perf_event *event, int flags)
|
|
{
|
|
struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events);
|
|
struct hw_perf_event *hwc = &event->hw;
|
|
int idx = hwc->idx;
|
|
|
|
if (__test_and_clear_bit(idx, cpuc->active_mask)) {
|
|
tile_pmu_disable_event(event);
|
|
cpuc->events[hwc->idx] = NULL;
|
|
WARN_ON_ONCE(hwc->state & PERF_HES_STOPPED);
|
|
hwc->state |= PERF_HES_STOPPED;
|
|
}
|
|
|
|
if ((flags & PERF_EF_UPDATE) && !(hwc->state & PERF_HES_UPTODATE)) {
|
|
/*
|
|
* Drain the remaining delta count out of a event
|
|
* that we are disabling:
|
|
*/
|
|
tile_perf_event_update(event);
|
|
hwc->state |= PERF_HES_UPTODATE;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Start an event (without re-assigning counter)
|
|
*/
|
|
static void tile_pmu_start(struct perf_event *event, int flags)
|
|
{
|
|
struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events);
|
|
int idx = event->hw.idx;
|
|
|
|
if (WARN_ON_ONCE(!(event->hw.state & PERF_HES_STOPPED)))
|
|
return;
|
|
|
|
if (WARN_ON_ONCE(idx == -1))
|
|
return;
|
|
|
|
if (flags & PERF_EF_RELOAD) {
|
|
WARN_ON_ONCE(!(event->hw.state & PERF_HES_UPTODATE));
|
|
tile_event_set_period(event);
|
|
}
|
|
|
|
event->hw.state = 0;
|
|
|
|
cpuc->events[idx] = event;
|
|
__set_bit(idx, cpuc->active_mask);
|
|
|
|
unmask_pmc_interrupts();
|
|
|
|
tile_pmu_enable_event(event);
|
|
|
|
perf_event_update_userpage(event);
|
|
}
|
|
|
|
/*
|
|
* Add a single event to the PMU.
|
|
*
|
|
* The event is added to the group of enabled events
|
|
* but only if it can be scehduled with existing events.
|
|
*/
|
|
static int tile_pmu_add(struct perf_event *event, int flags)
|
|
{
|
|
struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events);
|
|
struct hw_perf_event *hwc;
|
|
unsigned long mask;
|
|
int b, max_cnt;
|
|
|
|
hwc = &event->hw;
|
|
|
|
/*
|
|
* We are full.
|
|
*/
|
|
if (cpuc->n_events == tile_pmu->num_counters)
|
|
return -ENOSPC;
|
|
|
|
cpuc->event_list[cpuc->n_events] = event;
|
|
cpuc->n_events++;
|
|
|
|
hwc->state = PERF_HES_UPTODATE | PERF_HES_STOPPED;
|
|
if (!(flags & PERF_EF_START))
|
|
hwc->state |= PERF_HES_ARCH;
|
|
|
|
/*
|
|
* Find first empty counter.
|
|
*/
|
|
max_cnt = tile_pmu->num_counters;
|
|
mask = ~cpuc->used_mask;
|
|
|
|
/* Find next free counter. */
|
|
b = find_next_bit(&mask, max_cnt, 0);
|
|
|
|
/* Should not happen. */
|
|
if (WARN_ON_ONCE(b == max_cnt))
|
|
return -ENOSPC;
|
|
|
|
/*
|
|
* Assign counter to event.
|
|
*/
|
|
event->hw.idx = b;
|
|
__set_bit(b, &cpuc->used_mask);
|
|
|
|
/*
|
|
* Start if requested.
|
|
*/
|
|
if (flags & PERF_EF_START)
|
|
tile_pmu_start(event, PERF_EF_RELOAD);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Delete a single event from the PMU.
|
|
*
|
|
* The event is deleted from the group of enabled events.
|
|
* If it is the last event, disable PMU interrupt.
|
|
*/
|
|
static void tile_pmu_del(struct perf_event *event, int flags)
|
|
{
|
|
struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events);
|
|
int i;
|
|
|
|
/*
|
|
* Remove event from list, compact list if necessary.
|
|
*/
|
|
for (i = 0; i < cpuc->n_events; i++) {
|
|
if (cpuc->event_list[i] == event) {
|
|
while (++i < cpuc->n_events)
|
|
cpuc->event_list[i-1] = cpuc->event_list[i];
|
|
--cpuc->n_events;
|
|
cpuc->events[event->hw.idx] = NULL;
|
|
__clear_bit(event->hw.idx, &cpuc->used_mask);
|
|
tile_pmu_stop(event, PERF_EF_UPDATE);
|
|
break;
|
|
}
|
|
}
|
|
/*
|
|
* If there are no events left, then mask PMU interrupt.
|
|
*/
|
|
if (cpuc->n_events == 0)
|
|
mask_pmc_interrupts();
|
|
perf_event_update_userpage(event);
|
|
}
|
|
|
|
/*
|
|
* Propagate event elapsed time into the event.
|
|
*/
|
|
static inline void tile_pmu_read(struct perf_event *event)
|
|
{
|
|
tile_perf_event_update(event);
|
|
}
|
|
|
|
/*
|
|
* Map generic events to Tile PMU.
|
|
*/
|
|
static int tile_map_hw_event(u64 config)
|
|
{
|
|
if (config >= tile_pmu->max_events)
|
|
return -EINVAL;
|
|
return tile_pmu->hw_events[config];
|
|
}
|
|
|
|
/*
|
|
* Map generic hardware cache events to Tile PMU.
|
|
*/
|
|
static int tile_map_cache_event(u64 config)
|
|
{
|
|
unsigned int cache_type, cache_op, cache_result;
|
|
int code;
|
|
|
|
if (!tile_pmu->cache_events)
|
|
return -ENOENT;
|
|
|
|
cache_type = (config >> 0) & 0xff;
|
|
if (cache_type >= PERF_COUNT_HW_CACHE_MAX)
|
|
return -EINVAL;
|
|
|
|
cache_op = (config >> 8) & 0xff;
|
|
if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX)
|
|
return -EINVAL;
|
|
|
|
cache_result = (config >> 16) & 0xff;
|
|
if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX)
|
|
return -EINVAL;
|
|
|
|
code = (*tile_pmu->cache_events)[cache_type][cache_op][cache_result];
|
|
if (code == TILE_OP_UNSUPP)
|
|
return -EINVAL;
|
|
|
|
return code;
|
|
}
|
|
|
|
static void tile_event_destroy(struct perf_event *event)
|
|
{
|
|
if (atomic_dec_return(&tile_active_events) == 0)
|
|
release_pmc_hardware();
|
|
}
|
|
|
|
static int __tile_event_init(struct perf_event *event)
|
|
{
|
|
struct perf_event_attr *attr = &event->attr;
|
|
struct hw_perf_event *hwc = &event->hw;
|
|
int code;
|
|
|
|
switch (attr->type) {
|
|
case PERF_TYPE_HARDWARE:
|
|
code = tile_pmu->map_hw_event(attr->config);
|
|
break;
|
|
case PERF_TYPE_HW_CACHE:
|
|
code = tile_pmu->map_cache_event(attr->config);
|
|
break;
|
|
case PERF_TYPE_RAW:
|
|
code = attr->config & TILE_EVENT_MASK;
|
|
break;
|
|
default:
|
|
/* Should not happen. */
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
if (code < 0)
|
|
return code;
|
|
|
|
hwc->config = code;
|
|
hwc->idx = -1;
|
|
|
|
if (attr->exclude_user)
|
|
hwc->config |= TILE_CTL_EXCL_USER;
|
|
|
|
if (attr->exclude_kernel)
|
|
hwc->config |= TILE_CTL_EXCL_KERNEL;
|
|
|
|
if (attr->exclude_hv)
|
|
hwc->config |= TILE_CTL_EXCL_HV;
|
|
|
|
if (!hwc->sample_period) {
|
|
hwc->sample_period = tile_pmu->max_period;
|
|
hwc->last_period = hwc->sample_period;
|
|
local64_set(&hwc->period_left, hwc->sample_period);
|
|
}
|
|
event->destroy = tile_event_destroy;
|
|
return 0;
|
|
}
|
|
|
|
static int tile_event_init(struct perf_event *event)
|
|
{
|
|
int err = 0;
|
|
perf_irq_t old_irq_handler = NULL;
|
|
|
|
if (atomic_inc_return(&tile_active_events) == 1)
|
|
old_irq_handler = reserve_pmc_hardware(tile_pmu_handle_irq);
|
|
|
|
if (old_irq_handler) {
|
|
pr_warn("PMC hardware busy (reserved by oprofile)\n");
|
|
|
|
atomic_dec(&tile_active_events);
|
|
return -EBUSY;
|
|
}
|
|
|
|
switch (event->attr.type) {
|
|
case PERF_TYPE_RAW:
|
|
case PERF_TYPE_HARDWARE:
|
|
case PERF_TYPE_HW_CACHE:
|
|
break;
|
|
|
|
default:
|
|
return -ENOENT;
|
|
}
|
|
|
|
err = __tile_event_init(event);
|
|
if (err) {
|
|
if (event->destroy)
|
|
event->destroy(event);
|
|
}
|
|
return err;
|
|
}
|
|
|
|
static struct pmu tilera_pmu = {
|
|
.event_init = tile_event_init,
|
|
.add = tile_pmu_add,
|
|
.del = tile_pmu_del,
|
|
|
|
.start = tile_pmu_start,
|
|
.stop = tile_pmu_stop,
|
|
|
|
.read = tile_pmu_read,
|
|
};
|
|
|
|
/*
|
|
* PMU's IRQ handler, PMU has 2 interrupts, they share the same handler.
|
|
*/
|
|
int tile_pmu_handle_irq(struct pt_regs *regs, int fault)
|
|
{
|
|
struct perf_sample_data data;
|
|
struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events);
|
|
struct perf_event *event;
|
|
struct hw_perf_event *hwc;
|
|
u64 val;
|
|
unsigned long status;
|
|
int bit;
|
|
|
|
__this_cpu_inc(perf_irqs);
|
|
|
|
if (!atomic_read(&tile_active_events))
|
|
return 0;
|
|
|
|
status = pmc_get_overflow();
|
|
pmc_ack_overflow(status);
|
|
|
|
for_each_set_bit(bit, &status, tile_pmu->num_counters) {
|
|
|
|
event = cpuc->events[bit];
|
|
|
|
if (!event)
|
|
continue;
|
|
|
|
if (!test_bit(bit, cpuc->active_mask))
|
|
continue;
|
|
|
|
hwc = &event->hw;
|
|
|
|
val = tile_perf_event_update(event);
|
|
if (val & (1ULL << (tile_pmu->cntval_bits - 1)))
|
|
continue;
|
|
|
|
perf_sample_data_init(&data, 0, event->hw.last_period);
|
|
if (!tile_event_set_period(event))
|
|
continue;
|
|
|
|
if (perf_event_overflow(event, &data, regs))
|
|
tile_pmu_stop(event, 0);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static bool __init supported_pmu(void)
|
|
{
|
|
tile_pmu = &tilepmu;
|
|
return true;
|
|
}
|
|
|
|
int __init init_hw_perf_events(void)
|
|
{
|
|
supported_pmu();
|
|
perf_pmu_register(&tilera_pmu, "cpu", PERF_TYPE_RAW);
|
|
return 0;
|
|
}
|
|
arch_initcall(init_hw_perf_events);
|
|
|
|
/* Callchain handling code. */
|
|
|
|
/*
|
|
* Tile specific backtracing code for perf_events.
|
|
*/
|
|
static inline void perf_callchain(struct perf_callchain_entry *entry,
|
|
struct pt_regs *regs)
|
|
{
|
|
struct KBacktraceIterator kbt;
|
|
unsigned int i;
|
|
|
|
/*
|
|
* Get the address just after the "jalr" instruction that
|
|
* jumps to the handler for a syscall. When we find this
|
|
* address in a backtrace, we silently ignore it, which gives
|
|
* us a one-step backtrace connection from the sys_xxx()
|
|
* function in the kernel to the xxx() function in libc.
|
|
* Otherwise, we lose the ability to properly attribute time
|
|
* from the libc calls to the kernel implementations, since
|
|
* oprofile only considers PCs from backtraces a pair at a time.
|
|
*/
|
|
unsigned long handle_syscall_pc = handle_syscall_link_address();
|
|
|
|
KBacktraceIterator_init(&kbt, NULL, regs);
|
|
kbt.profile = 1;
|
|
|
|
/*
|
|
* The sample for the pc is already recorded. Now we are adding the
|
|
* address of the callsites on the stack. Our iterator starts
|
|
* with the frame of the (already sampled) call site. If our
|
|
* iterator contained a "return address" field, we could have just
|
|
* used it and wouldn't have needed to skip the first
|
|
* frame. That's in effect what the arm and x86 versions do.
|
|
* Instead we peel off the first iteration to get the equivalent
|
|
* behavior.
|
|
*/
|
|
|
|
if (KBacktraceIterator_end(&kbt))
|
|
return;
|
|
KBacktraceIterator_next(&kbt);
|
|
|
|
/*
|
|
* Set stack depth to 16 for user and kernel space respectively, that
|
|
* is, total 32 stack frames.
|
|
*/
|
|
for (i = 0; i < 16; ++i) {
|
|
unsigned long pc;
|
|
if (KBacktraceIterator_end(&kbt))
|
|
break;
|
|
pc = kbt.it.pc;
|
|
if (pc != handle_syscall_pc)
|
|
perf_callchain_store(entry, pc);
|
|
KBacktraceIterator_next(&kbt);
|
|
}
|
|
}
|
|
|
|
void perf_callchain_user(struct perf_callchain_entry *entry,
|
|
struct pt_regs *regs)
|
|
{
|
|
perf_callchain(entry, regs);
|
|
}
|
|
|
|
void perf_callchain_kernel(struct perf_callchain_entry *entry,
|
|
struct pt_regs *regs)
|
|
{
|
|
perf_callchain(entry, regs);
|
|
}
|