123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447 |
- /*
- * Xtensa Performance Monitor Module driver
- * See Tensilica Debug User's Guide for PMU registers documentation.
- *
- * Copyright (C) 2015 Cadence Design Systems Inc.
- *
- * This program is free software; you can redistribute it and/or modify
- * it under the terms of the GNU General Public License version 2 as
- * published by the Free Software Foundation.
- */
- #include <linux/interrupt.h>
- #include <linux/irqdomain.h>
- #include <linux/module.h>
- #include <linux/of.h>
- #include <linux/perf_event.h>
- #include <linux/platform_device.h>
- #include <asm/processor.h>
- #include <asm/stacktrace.h>
- /* Global control/status for all perf counters */
- #define XTENSA_PMU_PMG 0x1000
- /* Perf counter values */
- #define XTENSA_PMU_PM(i) (0x1080 + (i) * 4)
- /* Perf counter control registers */
- #define XTENSA_PMU_PMCTRL(i) (0x1100 + (i) * 4)
- /* Perf counter status registers */
- #define XTENSA_PMU_PMSTAT(i) (0x1180 + (i) * 4)
- #define XTENSA_PMU_PMG_PMEN 0x1
- #define XTENSA_PMU_COUNTER_MASK 0xffffffffULL
- #define XTENSA_PMU_COUNTER_MAX 0x7fffffff
- #define XTENSA_PMU_PMCTRL_INTEN 0x00000001
- #define XTENSA_PMU_PMCTRL_KRNLCNT 0x00000008
- #define XTENSA_PMU_PMCTRL_TRACELEVEL 0x000000f0
- #define XTENSA_PMU_PMCTRL_SELECT_SHIFT 8
- #define XTENSA_PMU_PMCTRL_SELECT 0x00001f00
- #define XTENSA_PMU_PMCTRL_MASK_SHIFT 16
- #define XTENSA_PMU_PMCTRL_MASK 0xffff0000
- #define XTENSA_PMU_MASK(select, mask) \
- (((select) << XTENSA_PMU_PMCTRL_SELECT_SHIFT) | \
- ((mask) << XTENSA_PMU_PMCTRL_MASK_SHIFT) | \
- XTENSA_PMU_PMCTRL_TRACELEVEL | \
- XTENSA_PMU_PMCTRL_INTEN)
- #define XTENSA_PMU_PMSTAT_OVFL 0x00000001
- #define XTENSA_PMU_PMSTAT_INTASRT 0x00000010
- struct xtensa_pmu_events {
- /* Array of events currently on this core */
- struct perf_event *event[XCHAL_NUM_PERF_COUNTERS];
- /* Bitmap of used hardware counters */
- unsigned long used_mask[BITS_TO_LONGS(XCHAL_NUM_PERF_COUNTERS)];
- };
- static DEFINE_PER_CPU(struct xtensa_pmu_events, xtensa_pmu_events);
- static const u32 xtensa_hw_ctl[] = {
- [PERF_COUNT_HW_CPU_CYCLES] = XTENSA_PMU_MASK(0, 0x1),
- [PERF_COUNT_HW_INSTRUCTIONS] = XTENSA_PMU_MASK(2, 0xffff),
- [PERF_COUNT_HW_CACHE_REFERENCES] = XTENSA_PMU_MASK(10, 0x1),
- [PERF_COUNT_HW_CACHE_MISSES] = XTENSA_PMU_MASK(12, 0x1),
- /* Taken and non-taken branches + taken loop ends */
- [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = XTENSA_PMU_MASK(2, 0x490),
- /* Instruction-related + other global stall cycles */
- [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = XTENSA_PMU_MASK(4, 0x1ff),
- /* Data-related global stall cycles */
- [PERF_COUNT_HW_STALLED_CYCLES_BACKEND] = XTENSA_PMU_MASK(3, 0x1ff),
- };
- #define C(_x) PERF_COUNT_HW_CACHE_##_x
- static const u32 xtensa_cache_ctl[][C(OP_MAX)][C(RESULT_MAX)] = {
- [C(L1D)] = {
- [C(OP_READ)] = {
- [C(RESULT_ACCESS)] = XTENSA_PMU_MASK(10, 0x1),
- [C(RESULT_MISS)] = XTENSA_PMU_MASK(10, 0x2),
- },
- [C(OP_WRITE)] = {
- [C(RESULT_ACCESS)] = XTENSA_PMU_MASK(11, 0x1),
- [C(RESULT_MISS)] = XTENSA_PMU_MASK(11, 0x2),
- },
- },
- [C(L1I)] = {
- [C(OP_READ)] = {
- [C(RESULT_ACCESS)] = XTENSA_PMU_MASK(8, 0x1),
- [C(RESULT_MISS)] = XTENSA_PMU_MASK(8, 0x2),
- },
- },
- [C(DTLB)] = {
- [C(OP_READ)] = {
- [C(RESULT_ACCESS)] = XTENSA_PMU_MASK(9, 0x1),
- [C(RESULT_MISS)] = XTENSA_PMU_MASK(9, 0x8),
- },
- },
- [C(ITLB)] = {
- [C(OP_READ)] = {
- [C(RESULT_ACCESS)] = XTENSA_PMU_MASK(7, 0x1),
- [C(RESULT_MISS)] = XTENSA_PMU_MASK(7, 0x8),
- },
- },
- };
- static int xtensa_pmu_cache_event(u64 config)
- {
- unsigned int cache_type, cache_op, cache_result;
- int ret;
- cache_type = (config >> 0) & 0xff;
- cache_op = (config >> 8) & 0xff;
- cache_result = (config >> 16) & 0xff;
- if (cache_type >= ARRAY_SIZE(xtensa_cache_ctl) ||
- cache_op >= C(OP_MAX) ||
- cache_result >= C(RESULT_MAX))
- return -EINVAL;
- ret = xtensa_cache_ctl[cache_type][cache_op][cache_result];
- if (ret == 0)
- return -EINVAL;
- return ret;
- }
- static inline uint32_t xtensa_pmu_read_counter(int idx)
- {
- return get_er(XTENSA_PMU_PM(idx));
- }
- static inline void xtensa_pmu_write_counter(int idx, uint32_t v)
- {
- set_er(v, XTENSA_PMU_PM(idx));
- }
- static void xtensa_perf_event_update(struct perf_event *event,
- struct hw_perf_event *hwc, int idx)
- {
- uint64_t prev_raw_count, new_raw_count;
- int64_t delta;
- do {
- prev_raw_count = local64_read(&hwc->prev_count);
- new_raw_count = xtensa_pmu_read_counter(event->hw.idx);
- } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count,
- new_raw_count) != prev_raw_count);
- delta = (new_raw_count - prev_raw_count) & XTENSA_PMU_COUNTER_MASK;
- local64_add(delta, &event->count);
- local64_sub(delta, &hwc->period_left);
- }
- static bool xtensa_perf_event_set_period(struct perf_event *event,
- struct hw_perf_event *hwc, int idx)
- {
- bool rc = false;
- s64 left;
- if (!is_sampling_event(event)) {
- left = XTENSA_PMU_COUNTER_MAX;
- } else {
- s64 period = hwc->sample_period;
- left = local64_read(&hwc->period_left);
- if (left <= -period) {
- left = period;
- local64_set(&hwc->period_left, left);
- hwc->last_period = period;
- rc = true;
- } else if (left <= 0) {
- left += period;
- local64_set(&hwc->period_left, left);
- hwc->last_period = period;
- rc = true;
- }
- if (left > XTENSA_PMU_COUNTER_MAX)
- left = XTENSA_PMU_COUNTER_MAX;
- }
- local64_set(&hwc->prev_count, -left);
- xtensa_pmu_write_counter(idx, -left);
- perf_event_update_userpage(event);
- return rc;
- }
- static void xtensa_pmu_enable(struct pmu *pmu)
- {
- set_er(get_er(XTENSA_PMU_PMG) | XTENSA_PMU_PMG_PMEN, XTENSA_PMU_PMG);
- }
- static void xtensa_pmu_disable(struct pmu *pmu)
- {
- set_er(get_er(XTENSA_PMU_PMG) & ~XTENSA_PMU_PMG_PMEN, XTENSA_PMU_PMG);
- }
- static int xtensa_pmu_event_init(struct perf_event *event)
- {
- int ret;
- switch (event->attr.type) {
- case PERF_TYPE_HARDWARE:
- if (event->attr.config >= ARRAY_SIZE(xtensa_hw_ctl) ||
- xtensa_hw_ctl[event->attr.config] == 0)
- return -EINVAL;
- event->hw.config = xtensa_hw_ctl[event->attr.config];
- return 0;
- case PERF_TYPE_HW_CACHE:
- ret = xtensa_pmu_cache_event(event->attr.config);
- if (ret < 0)
- return ret;
- event->hw.config = ret;
- return 0;
- case PERF_TYPE_RAW:
- /* Not 'previous counter' select */
- if ((event->attr.config & XTENSA_PMU_PMCTRL_SELECT) ==
- (1 << XTENSA_PMU_PMCTRL_SELECT_SHIFT))
- return -EINVAL;
- event->hw.config = (event->attr.config &
- (XTENSA_PMU_PMCTRL_KRNLCNT |
- XTENSA_PMU_PMCTRL_TRACELEVEL |
- XTENSA_PMU_PMCTRL_SELECT |
- XTENSA_PMU_PMCTRL_MASK)) |
- XTENSA_PMU_PMCTRL_INTEN;
- return 0;
- default:
- return -ENOENT;
- }
- }
- /*
- * Starts/Stops a counter present on the PMU. The PMI handler
- * should stop the counter when perf_event_overflow() returns
- * !0. ->start() will be used to continue.
- */
- static void xtensa_pmu_start(struct perf_event *event, int flags)
- {
- struct hw_perf_event *hwc = &event->hw;
- int idx = hwc->idx;
- if (WARN_ON_ONCE(idx == -1))
- return;
- if (flags & PERF_EF_RELOAD) {
- WARN_ON_ONCE(!(event->hw.state & PERF_HES_UPTODATE));
- xtensa_perf_event_set_period(event, hwc, idx);
- }
- hwc->state = 0;
- set_er(hwc->config, XTENSA_PMU_PMCTRL(idx));
- }
- static void xtensa_pmu_stop(struct perf_event *event, int flags)
- {
- struct hw_perf_event *hwc = &event->hw;
- int idx = hwc->idx;
- if (!(hwc->state & PERF_HES_STOPPED)) {
- set_er(0, XTENSA_PMU_PMCTRL(idx));
- set_er(get_er(XTENSA_PMU_PMSTAT(idx)),
- XTENSA_PMU_PMSTAT(idx));
- hwc->state |= PERF_HES_STOPPED;
- }
- if ((flags & PERF_EF_UPDATE) &&
- !(event->hw.state & PERF_HES_UPTODATE)) {
- xtensa_perf_event_update(event, &event->hw, idx);
- event->hw.state |= PERF_HES_UPTODATE;
- }
- }
- /*
- * Adds/Removes a counter to/from the PMU, can be done inside
- * a transaction, see the ->*_txn() methods.
- */
- static int xtensa_pmu_add(struct perf_event *event, int flags)
- {
- struct xtensa_pmu_events *ev = this_cpu_ptr(&xtensa_pmu_events);
- struct hw_perf_event *hwc = &event->hw;
- int idx = hwc->idx;
- if (__test_and_set_bit(idx, ev->used_mask)) {
- idx = find_first_zero_bit(ev->used_mask,
- XCHAL_NUM_PERF_COUNTERS);
- if (idx == XCHAL_NUM_PERF_COUNTERS)
- return -EAGAIN;
- __set_bit(idx, ev->used_mask);
- hwc->idx = idx;
- }
- ev->event[idx] = event;
- hwc->state = PERF_HES_UPTODATE | PERF_HES_STOPPED;
- if (flags & PERF_EF_START)
- xtensa_pmu_start(event, PERF_EF_RELOAD);
- perf_event_update_userpage(event);
- return 0;
- }
- static void xtensa_pmu_del(struct perf_event *event, int flags)
- {
- struct xtensa_pmu_events *ev = this_cpu_ptr(&xtensa_pmu_events);
- xtensa_pmu_stop(event, PERF_EF_UPDATE);
- __clear_bit(event->hw.idx, ev->used_mask);
- perf_event_update_userpage(event);
- }
- static void xtensa_pmu_read(struct perf_event *event)
- {
- xtensa_perf_event_update(event, &event->hw, event->hw.idx);
- }
- static int callchain_trace(struct stackframe *frame, void *data)
- {
- struct perf_callchain_entry_ctx *entry = data;
- perf_callchain_store(entry, frame->pc);
- return 0;
- }
- void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry,
- struct pt_regs *regs)
- {
- xtensa_backtrace_kernel(regs, entry->max_stack,
- callchain_trace, NULL, entry);
- }
- void perf_callchain_user(struct perf_callchain_entry_ctx *entry,
- struct pt_regs *regs)
- {
- xtensa_backtrace_user(regs, entry->max_stack,
- callchain_trace, entry);
- }
- void perf_event_print_debug(void)
- {
- unsigned long flags;
- unsigned i;
- local_irq_save(flags);
- pr_info("CPU#%d: PMG: 0x%08lx\n", smp_processor_id(),
- get_er(XTENSA_PMU_PMG));
- for (i = 0; i < XCHAL_NUM_PERF_COUNTERS; ++i)
- pr_info("PM%d: 0x%08lx, PMCTRL%d: 0x%08lx, PMSTAT%d: 0x%08lx\n",
- i, get_er(XTENSA_PMU_PM(i)),
- i, get_er(XTENSA_PMU_PMCTRL(i)),
- i, get_er(XTENSA_PMU_PMSTAT(i)));
- local_irq_restore(flags);
- }
- irqreturn_t xtensa_pmu_irq_handler(int irq, void *dev_id)
- {
- irqreturn_t rc = IRQ_NONE;
- struct xtensa_pmu_events *ev = this_cpu_ptr(&xtensa_pmu_events);
- unsigned i;
- for (i = find_first_bit(ev->used_mask, XCHAL_NUM_PERF_COUNTERS);
- i < XCHAL_NUM_PERF_COUNTERS;
- i = find_next_bit(ev->used_mask, XCHAL_NUM_PERF_COUNTERS, i + 1)) {
- uint32_t v = get_er(XTENSA_PMU_PMSTAT(i));
- struct perf_event *event = ev->event[i];
- struct hw_perf_event *hwc = &event->hw;
- u64 last_period;
- if (!(v & XTENSA_PMU_PMSTAT_OVFL))
- continue;
- set_er(v, XTENSA_PMU_PMSTAT(i));
- xtensa_perf_event_update(event, hwc, i);
- last_period = hwc->last_period;
- if (xtensa_perf_event_set_period(event, hwc, i)) {
- struct perf_sample_data data;
- struct pt_regs *regs = get_irq_regs();
- perf_sample_data_init(&data, 0, last_period);
- if (perf_event_overflow(event, &data, regs))
- xtensa_pmu_stop(event, 0);
- }
- rc = IRQ_HANDLED;
- }
- return rc;
- }
- static struct pmu xtensa_pmu = {
- .pmu_enable = xtensa_pmu_enable,
- .pmu_disable = xtensa_pmu_disable,
- .event_init = xtensa_pmu_event_init,
- .add = xtensa_pmu_add,
- .del = xtensa_pmu_del,
- .start = xtensa_pmu_start,
- .stop = xtensa_pmu_stop,
- .read = xtensa_pmu_read,
- };
- static int xtensa_pmu_setup(int cpu)
- {
- unsigned i;
- set_er(0, XTENSA_PMU_PMG);
- for (i = 0; i < XCHAL_NUM_PERF_COUNTERS; ++i) {
- set_er(0, XTENSA_PMU_PMCTRL(i));
- set_er(get_er(XTENSA_PMU_PMSTAT(i)), XTENSA_PMU_PMSTAT(i));
- }
- return 0;
- }
- static int __init xtensa_pmu_init(void)
- {
- int ret;
- int irq = irq_create_mapping(NULL, XCHAL_PROFILING_INTERRUPT);
- ret = cpuhp_setup_state(CPUHP_AP_PERF_XTENSA_STARTING,
- "AP_PERF_XTENSA_STARTING", xtensa_pmu_setup,
- NULL);
- if (ret) {
- pr_err("xtensa_pmu: failed to register CPU-hotplug.\n");
- return ret;
- }
- #if XTENSA_FAKE_NMI
- enable_irq(irq);
- #else
- ret = request_irq(irq, xtensa_pmu_irq_handler, IRQF_PERCPU,
- "pmu", NULL);
- if (ret < 0)
- return ret;
- #endif
- ret = perf_pmu_register(&xtensa_pmu, "cpu", PERF_TYPE_RAW);
- if (ret)
- free_irq(irq, NULL);
- return ret;
- }
- early_initcall(xtensa_pmu_init);
|