linux/arch/riscv/kernel/perf_event.c
<<
>>
Prefs
   1/* SPDX-License-Identifier: GPL-2.0 */
   2/*
   3 * Copyright (C) 2008 Thomas Gleixner <tglx@linutronix.de>
   4 * Copyright (C) 2008-2009 Red Hat, Inc., Ingo Molnar
   5 * Copyright (C) 2009 Jaswinder Singh Rajput
   6 * Copyright (C) 2009 Advanced Micro Devices, Inc., Robert Richter
   7 * Copyright (C) 2008-2009 Red Hat, Inc., Peter Zijlstra
   8 * Copyright (C) 2009 Intel Corporation, <markus.t.metzger@intel.com>
   9 * Copyright (C) 2009 Google, Inc., Stephane Eranian
  10 * Copyright 2014 Tilera Corporation. All Rights Reserved.
  11 * Copyright (C) 2018 Andes Technology Corporation
  12 *
  13 * Perf_events support for RISC-V platforms.
  14 *
  15 * Since the spec. (as of now, Priv-Spec 1.10) does not provide enough
  16 * functionality for perf event to fully work, this file provides
  17 * the very basic framework only.
  18 *
  19 * For platform portings, please check Documentations/riscv/pmu.txt.
  20 *
  21 * The Copyright line includes x86 and tile ones.
  22 */
  23
  24#include <linux/kprobes.h>
  25#include <linux/kernel.h>
  26#include <linux/kdebug.h>
  27#include <linux/mutex.h>
  28#include <linux/bitmap.h>
  29#include <linux/irq.h>
  30#include <linux/perf_event.h>
  31#include <linux/atomic.h>
  32#include <linux/of.h>
  33#include <asm/perf_event.h>
  34
  35static const struct riscv_pmu *riscv_pmu __read_mostly;
  36static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events);
  37
  38/*
  39 * Hardware & cache maps and their methods
  40 */
  41
  42static const int riscv_hw_event_map[] = {
  43        [PERF_COUNT_HW_CPU_CYCLES]              = RISCV_PMU_CYCLE,
  44        [PERF_COUNT_HW_INSTRUCTIONS]            = RISCV_PMU_INSTRET,
  45        [PERF_COUNT_HW_CACHE_REFERENCES]        = RISCV_OP_UNSUPP,
  46        [PERF_COUNT_HW_CACHE_MISSES]            = RISCV_OP_UNSUPP,
  47        [PERF_COUNT_HW_BRANCH_INSTRUCTIONS]     = RISCV_OP_UNSUPP,
  48        [PERF_COUNT_HW_BRANCH_MISSES]           = RISCV_OP_UNSUPP,
  49        [PERF_COUNT_HW_BUS_CYCLES]              = RISCV_OP_UNSUPP,
  50};
  51
  52#define C(x) PERF_COUNT_HW_CACHE_##x
  53static const int riscv_cache_event_map[PERF_COUNT_HW_CACHE_MAX]
  54[PERF_COUNT_HW_CACHE_OP_MAX]
  55[PERF_COUNT_HW_CACHE_RESULT_MAX] = {
  56        [C(L1D)] = {
  57                [C(OP_READ)] = {
  58                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
  59                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
  60                },
  61                [C(OP_WRITE)] = {
  62                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
  63                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
  64                },
  65                [C(OP_PREFETCH)] = {
  66                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
  67                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
  68                },
  69        },
  70        [C(L1I)] = {
  71                [C(OP_READ)] = {
  72                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
  73                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
  74                },
  75                [C(OP_WRITE)] = {
  76                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
  77                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
  78                },
  79                [C(OP_PREFETCH)] = {
  80                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
  81                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
  82                },
  83        },
  84        [C(LL)] = {
  85                [C(OP_READ)] = {
  86                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
  87                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
  88                },
  89                [C(OP_WRITE)] = {
  90                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
  91                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
  92                },
  93                [C(OP_PREFETCH)] = {
  94                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
  95                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
  96                },
  97        },
  98        [C(DTLB)] = {
  99                [C(OP_READ)] = {
 100                        [C(RESULT_ACCESS)] =  RISCV_OP_UNSUPP,
 101                        [C(RESULT_MISS)] =  RISCV_OP_UNSUPP,
 102                },
 103                [C(OP_WRITE)] = {
 104                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
 105                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
 106                },
 107                [C(OP_PREFETCH)] = {
 108                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
 109                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
 110                },
 111        },
 112        [C(ITLB)] = {
 113                [C(OP_READ)] = {
 114                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
 115                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
 116                },
 117                [C(OP_WRITE)] = {
 118                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
 119                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
 120                },
 121                [C(OP_PREFETCH)] = {
 122                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
 123                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
 124                },
 125        },
 126        [C(BPU)] = {
 127                [C(OP_READ)] = {
 128                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
 129                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
 130                },
 131                [C(OP_WRITE)] = {
 132                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
 133                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
 134                },
 135                [C(OP_PREFETCH)] = {
 136                        [C(RESULT_ACCESS)] = RISCV_OP_UNSUPP,
 137                        [C(RESULT_MISS)] = RISCV_OP_UNSUPP,
 138                },
 139        },
 140};
 141
 142static int riscv_map_hw_event(u64 config)
 143{
 144        if (config >= riscv_pmu->max_events)
 145                return -EINVAL;
 146
 147        return riscv_pmu->hw_events[config];
 148}
 149
 150static int riscv_map_cache_decode(u64 config, unsigned int *type,
 151                           unsigned int *op, unsigned int *result)
 152{
 153        return -ENOENT;
 154}
 155
 156static int riscv_map_cache_event(u64 config)
 157{
 158        unsigned int type, op, result;
 159        int err = -ENOENT;
 160                int code;
 161
 162        err = riscv_map_cache_decode(config, &type, &op, &result);
 163        if (!riscv_pmu->cache_events || err)
 164                return err;
 165
 166        if (type >= PERF_COUNT_HW_CACHE_MAX ||
 167            op >= PERF_COUNT_HW_CACHE_OP_MAX ||
 168            result >= PERF_COUNT_HW_CACHE_RESULT_MAX)
 169                return -EINVAL;
 170
 171        code = (*riscv_pmu->cache_events)[type][op][result];
 172        if (code == RISCV_OP_UNSUPP)
 173                return -EINVAL;
 174
 175        return code;
 176}
 177
 178/*
 179 * Low-level functions: reading/writing counters
 180 */
 181
 182static inline u64 read_counter(int idx)
 183{
 184        u64 val = 0;
 185
 186        switch (idx) {
 187        case RISCV_PMU_CYCLE:
 188                val = csr_read(CSR_CYCLE);
 189                break;
 190        case RISCV_PMU_INSTRET:
 191                val = csr_read(CSR_INSTRET);
 192                break;
 193        default:
 194                WARN_ON_ONCE(idx < 0 || idx > RISCV_MAX_COUNTERS);
 195                return -EINVAL;
 196        }
 197
 198        return val;
 199}
 200
 201static inline void write_counter(int idx, u64 value)
 202{
 203        /* currently not supported */
 204        WARN_ON_ONCE(1);
 205}
 206
 207/*
 208 * pmu->read: read and update the counter
 209 *
 210 * Other architectures' implementation often have a xxx_perf_event_update
 211 * routine, which can return counter values when called in the IRQ, but
 212 * return void when being called by the pmu->read method.
 213 */
 214static void riscv_pmu_read(struct perf_event *event)
 215{
 216        struct hw_perf_event *hwc = &event->hw;
 217        u64 prev_raw_count, new_raw_count;
 218        u64 oldval;
 219        int idx = hwc->idx;
 220        u64 delta;
 221
 222        do {
 223                prev_raw_count = local64_read(&hwc->prev_count);
 224                new_raw_count = read_counter(idx);
 225
 226                oldval = local64_cmpxchg(&hwc->prev_count, prev_raw_count,
 227                                         new_raw_count);
 228        } while (oldval != prev_raw_count);
 229
 230        /*
 231         * delta is the value to update the counter we maintain in the kernel.
 232         */
 233        delta = (new_raw_count - prev_raw_count) &
 234                ((1ULL << riscv_pmu->counter_width) - 1);
 235        local64_add(delta, &event->count);
 236        /*
 237         * Something like local64_sub(delta, &hwc->period_left) here is
 238         * needed if there is an interrupt for perf.
 239         */
 240}
 241
 242/*
 243 * State transition functions:
 244 *
 245 * stop()/start() & add()/del()
 246 */
 247
 248/*
 249 * pmu->stop: stop the counter
 250 */
 251static void riscv_pmu_stop(struct perf_event *event, int flags)
 252{
 253        struct hw_perf_event *hwc = &event->hw;
 254
 255        WARN_ON_ONCE(hwc->state & PERF_HES_STOPPED);
 256        hwc->state |= PERF_HES_STOPPED;
 257
 258        if ((flags & PERF_EF_UPDATE) && !(hwc->state & PERF_HES_UPTODATE)) {
 259                riscv_pmu->pmu->read(event);
 260                hwc->state |= PERF_HES_UPTODATE;
 261        }
 262}
 263
 264/*
 265 * pmu->start: start the event.
 266 */
 267static void riscv_pmu_start(struct perf_event *event, int flags)
 268{
 269        struct hw_perf_event *hwc = &event->hw;
 270
 271        if (WARN_ON_ONCE(!(event->hw.state & PERF_HES_STOPPED)))
 272                return;
 273
 274        if (flags & PERF_EF_RELOAD) {
 275                WARN_ON_ONCE(!(event->hw.state & PERF_HES_UPTODATE));
 276
 277                /*
 278                 * Set the counter to the period to the next interrupt here,
 279                 * if you have any.
 280                 */
 281        }
 282
 283        hwc->state = 0;
 284        perf_event_update_userpage(event);
 285
 286        /*
 287         * Since we cannot write to counters, this serves as an initialization
 288         * to the delta-mechanism in pmu->read(); otherwise, the delta would be
 289         * wrong when pmu->read is called for the first time.
 290         */
 291        local64_set(&hwc->prev_count, read_counter(hwc->idx));
 292}
 293
 294/*
 295 * pmu->add: add the event to PMU.
 296 */
 297static int riscv_pmu_add(struct perf_event *event, int flags)
 298{
 299        struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events);
 300        struct hw_perf_event *hwc = &event->hw;
 301
 302        if (cpuc->n_events == riscv_pmu->num_counters)
 303                return -ENOSPC;
 304
 305        /*
 306         * We don't have general conunters, so no binding-event-to-counter
 307         * process here.
 308         *
 309         * Indexing using hwc->config generally not works, since config may
 310         * contain extra information, but here the only info we have in
 311         * hwc->config is the event index.
 312         */
 313        hwc->idx = hwc->config;
 314        cpuc->events[hwc->idx] = event;
 315        cpuc->n_events++;
 316
 317        hwc->state = PERF_HES_UPTODATE | PERF_HES_STOPPED;
 318
 319        if (flags & PERF_EF_START)
 320                riscv_pmu->pmu->start(event, PERF_EF_RELOAD);
 321
 322        return 0;
 323}
 324
 325/*
 326 * pmu->del: delete the event from PMU.
 327 */
 328static void riscv_pmu_del(struct perf_event *event, int flags)
 329{
 330        struct cpu_hw_events *cpuc = this_cpu_ptr(&cpu_hw_events);
 331        struct hw_perf_event *hwc = &event->hw;
 332
 333        cpuc->events[hwc->idx] = NULL;
 334        cpuc->n_events--;
 335        riscv_pmu->pmu->stop(event, PERF_EF_UPDATE);
 336        perf_event_update_userpage(event);
 337}
 338
 339/*
 340 * Interrupt: a skeletion for reference.
 341 */
 342
 343static DEFINE_MUTEX(pmc_reserve_mutex);
 344
 345static irqreturn_t riscv_base_pmu_handle_irq(int irq_num, void *dev)
 346{
 347        return IRQ_NONE;
 348}
 349
 350static int reserve_pmc_hardware(void)
 351{
 352        int err = 0;
 353
 354        mutex_lock(&pmc_reserve_mutex);
 355        if (riscv_pmu->irq >= 0 && riscv_pmu->handle_irq) {
 356                err = request_irq(riscv_pmu->irq, riscv_pmu->handle_irq,
 357                                  IRQF_PERCPU, "riscv-base-perf", NULL);
 358        }
 359        mutex_unlock(&pmc_reserve_mutex);
 360
 361        return err;
 362}
 363
 364static void release_pmc_hardware(void)
 365{
 366        mutex_lock(&pmc_reserve_mutex);
 367        if (riscv_pmu->irq >= 0)
 368                free_irq(riscv_pmu->irq, NULL);
 369        mutex_unlock(&pmc_reserve_mutex);
 370}
 371
 372/*
 373 * Event Initialization/Finalization
 374 */
 375
 376static atomic_t riscv_active_events = ATOMIC_INIT(0);
 377
 378static void riscv_event_destroy(struct perf_event *event)
 379{
 380        if (atomic_dec_return(&riscv_active_events) == 0)
 381                release_pmc_hardware();
 382}
 383
 384static int riscv_event_init(struct perf_event *event)
 385{
 386        struct perf_event_attr *attr = &event->attr;
 387        struct hw_perf_event *hwc = &event->hw;
 388        int err;
 389        int code;
 390
 391        if (atomic_inc_return(&riscv_active_events) == 1) {
 392                err = reserve_pmc_hardware();
 393
 394                if (err) {
 395                        pr_warn("PMC hardware not available\n");
 396                        atomic_dec(&riscv_active_events);
 397                        return -EBUSY;
 398                }
 399        }
 400
 401        switch (event->attr.type) {
 402        case PERF_TYPE_HARDWARE:
 403                code = riscv_pmu->map_hw_event(attr->config);
 404                break;
 405        case PERF_TYPE_HW_CACHE:
 406                code = riscv_pmu->map_cache_event(attr->config);
 407                break;
 408        case PERF_TYPE_RAW:
 409                return -EOPNOTSUPP;
 410        default:
 411                return -ENOENT;
 412        }
 413
 414        event->destroy = riscv_event_destroy;
 415        if (code < 0) {
 416                event->destroy(event);
 417                return code;
 418        }
 419
 420        /*
 421         * idx is set to -1 because the index of a general event should not be
 422         * decided until binding to some counter in pmu->add().
 423         *
 424         * But since we don't have such support, later in pmu->add(), we just
 425         * use hwc->config as the index instead.
 426         */
 427        hwc->config = code;
 428        hwc->idx = -1;
 429
 430        return 0;
 431}
 432
 433/*
 434 * Initialization
 435 */
 436
 437static struct pmu min_pmu = {
 438        .name           = "riscv-base",
 439        .event_init     = riscv_event_init,
 440        .add            = riscv_pmu_add,
 441        .del            = riscv_pmu_del,
 442        .start          = riscv_pmu_start,
 443        .stop           = riscv_pmu_stop,
 444        .read           = riscv_pmu_read,
 445};
 446
 447static const struct riscv_pmu riscv_base_pmu = {
 448        .pmu = &min_pmu,
 449        .max_events = ARRAY_SIZE(riscv_hw_event_map),
 450        .map_hw_event = riscv_map_hw_event,
 451        .hw_events = riscv_hw_event_map,
 452        .map_cache_event = riscv_map_cache_event,
 453        .cache_events = &riscv_cache_event_map,
 454        .counter_width = 63,
 455        .num_counters = RISCV_BASE_COUNTERS + 0,
 456        .handle_irq = &riscv_base_pmu_handle_irq,
 457
 458        /* This means this PMU has no IRQ. */
 459        .irq = -1,
 460};
 461
 462static const struct of_device_id riscv_pmu_of_ids[] = {
 463        {.compatible = "riscv,base-pmu",        .data = &riscv_base_pmu},
 464        { /* sentinel value */ }
 465};
 466
 467static int __init init_hw_perf_events(void)
 468{
 469        struct device_node *node = of_find_node_by_type(NULL, "pmu");
 470        const struct of_device_id *of_id;
 471
 472        riscv_pmu = &riscv_base_pmu;
 473
 474        if (node) {
 475                of_id = of_match_node(riscv_pmu_of_ids, node);
 476
 477                if (of_id)
 478                        riscv_pmu = of_id->data;
 479                of_node_put(node);
 480        }
 481
 482        perf_pmu_register(riscv_pmu->pmu, "cpu", PERF_TYPE_RAW);
 483        return 0;
 484}
 485arch_initcall(init_hw_perf_events);
 486