linux/tools/bpf/bpftool/map_perf_ring.c
<<
>>
Prefs
   1// SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause)
   2/* Copyright (C) 2018 Netronome Systems, Inc. */
   3/* This program is free software; you can redistribute it and/or
   4 * modify it under the terms of version 2 of the GNU General Public
   5 * License as published by the Free Software Foundation.
   6 */
   7#include <errno.h>
   8#include <fcntl.h>
   9#include <libbpf.h>
  10#include <poll.h>
  11#include <signal.h>
  12#include <stdbool.h>
  13#include <stdio.h>
  14#include <stdlib.h>
  15#include <string.h>
  16#include <time.h>
  17#include <unistd.h>
  18#include <linux/bpf.h>
  19#include <linux/perf_event.h>
  20#include <sys/ioctl.h>
  21#include <sys/mman.h>
  22#include <sys/syscall.h>
  23
  24#include <bpf.h>
  25#include <perf-sys.h>
  26
  27#include "main.h"
  28
  29#define MMAP_PAGE_CNT   16
  30
  31static bool stop;
  32
  33struct event_ring_info {
  34        int fd;
  35        int key;
  36        unsigned int cpu;
  37        void *mem;
  38};
  39
  40struct perf_event_sample {
  41        struct perf_event_header header;
  42        u64 time;
  43        __u32 size;
  44        unsigned char data[];
  45};
  46
  47static void int_exit(int signo)
  48{
  49        fprintf(stderr, "Stopping...\n");
  50        stop = true;
  51}
  52
  53static enum bpf_perf_event_ret
  54print_bpf_output(struct perf_event_header *event, void *private_data)
  55{
  56        struct perf_event_sample *e = container_of(event, struct perf_event_sample,
  57                                                   header);
  58        struct event_ring_info *ring = private_data;
  59        struct {
  60                struct perf_event_header header;
  61                __u64 id;
  62                __u64 lost;
  63        } *lost = (typeof(lost))event;
  64
  65        if (json_output) {
  66                jsonw_start_object(json_wtr);
  67                jsonw_name(json_wtr, "type");
  68                jsonw_uint(json_wtr, e->header.type);
  69                jsonw_name(json_wtr, "cpu");
  70                jsonw_uint(json_wtr, ring->cpu);
  71                jsonw_name(json_wtr, "index");
  72                jsonw_uint(json_wtr, ring->key);
  73                if (e->header.type == PERF_RECORD_SAMPLE) {
  74                        jsonw_name(json_wtr, "timestamp");
  75                        jsonw_uint(json_wtr, e->time);
  76                        jsonw_name(json_wtr, "data");
  77                        print_data_json(e->data, e->size);
  78                } else if (e->header.type == PERF_RECORD_LOST) {
  79                        jsonw_name(json_wtr, "lost");
  80                        jsonw_start_object(json_wtr);
  81                        jsonw_name(json_wtr, "id");
  82                        jsonw_uint(json_wtr, lost->id);
  83                        jsonw_name(json_wtr, "count");
  84                        jsonw_uint(json_wtr, lost->lost);
  85                        jsonw_end_object(json_wtr);
  86                }
  87                jsonw_end_object(json_wtr);
  88        } else {
  89                if (e->header.type == PERF_RECORD_SAMPLE) {
  90                        printf("== @%lld.%09lld CPU: %d index: %d =====\n",
  91                               e->time / 1000000000ULL, e->time % 1000000000ULL,
  92                               ring->cpu, ring->key);
  93                        fprint_hex(stdout, e->data, e->size, " ");
  94                        printf("\n");
  95                } else if (e->header.type == PERF_RECORD_LOST) {
  96                        printf("lost %lld events\n", lost->lost);
  97                } else {
  98                        printf("unknown event type=%d size=%d\n",
  99                               e->header.type, e->header.size);
 100                }
 101        }
 102
 103        return LIBBPF_PERF_EVENT_CONT;
 104}
 105
 106static void
 107perf_event_read(struct event_ring_info *ring, void **buf, size_t *buf_len)
 108{
 109        enum bpf_perf_event_ret ret;
 110
 111        ret = bpf_perf_event_read_simple(ring->mem,
 112                                         MMAP_PAGE_CNT * get_page_size(),
 113                                         get_page_size(), buf, buf_len,
 114                                         print_bpf_output, ring);
 115        if (ret != LIBBPF_PERF_EVENT_CONT) {
 116                fprintf(stderr, "perf read loop failed with %d\n", ret);
 117                stop = true;
 118        }
 119}
 120
 121static int perf_mmap_size(void)
 122{
 123        return get_page_size() * (MMAP_PAGE_CNT + 1);
 124}
 125
 126static void *perf_event_mmap(int fd)
 127{
 128        int mmap_size = perf_mmap_size();
 129        void *base;
 130
 131        base = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
 132        if (base == MAP_FAILED) {
 133                p_err("event mmap failed: %s\n", strerror(errno));
 134                return NULL;
 135        }
 136
 137        return base;
 138}
 139
 140static void perf_event_unmap(void *mem)
 141{
 142        if (munmap(mem, perf_mmap_size()))
 143                fprintf(stderr, "Can't unmap ring memory!\n");
 144}
 145
 146static int bpf_perf_event_open(int map_fd, int key, int cpu)
 147{
 148        struct perf_event_attr attr = {
 149                .sample_type = PERF_SAMPLE_RAW | PERF_SAMPLE_TIME,
 150                .type = PERF_TYPE_SOFTWARE,
 151                .config = PERF_COUNT_SW_BPF_OUTPUT,
 152        };
 153        int pmu_fd;
 154
 155        pmu_fd = sys_perf_event_open(&attr, -1, cpu, -1, 0);
 156        if (pmu_fd < 0) {
 157                p_err("failed to open perf event %d for CPU %d", key, cpu);
 158                return -1;
 159        }
 160
 161        if (bpf_map_update_elem(map_fd, &key, &pmu_fd, BPF_ANY)) {
 162                p_err("failed to update map for event %d for CPU %d", key, cpu);
 163                goto err_close;
 164        }
 165        if (ioctl(pmu_fd, PERF_EVENT_IOC_ENABLE, 0)) {
 166                p_err("failed to enable event %d for CPU %d", key, cpu);
 167                goto err_close;
 168        }
 169
 170        return pmu_fd;
 171
 172err_close:
 173        close(pmu_fd);
 174        return -1;
 175}
 176
 177int do_event_pipe(int argc, char **argv)
 178{
 179        int i, nfds, map_fd, index = -1, cpu = -1;
 180        struct bpf_map_info map_info = {};
 181        struct event_ring_info *rings;
 182        size_t tmp_buf_sz = 0;
 183        void *tmp_buf = NULL;
 184        struct pollfd *pfds;
 185        __u32 map_info_len;
 186        bool do_all = true;
 187
 188        map_info_len = sizeof(map_info);
 189        map_fd = map_parse_fd_and_info(&argc, &argv, &map_info, &map_info_len);
 190        if (map_fd < 0)
 191                return -1;
 192
 193        if (map_info.type != BPF_MAP_TYPE_PERF_EVENT_ARRAY) {
 194                p_err("map is not a perf event array");
 195                goto err_close_map;
 196        }
 197
 198        while (argc) {
 199                if (argc < 2) {
 200                        BAD_ARG();
 201                        goto err_close_map;
 202                }
 203
 204                if (is_prefix(*argv, "cpu")) {
 205                        char *endptr;
 206
 207                        NEXT_ARG();
 208                        cpu = strtoul(*argv, &endptr, 0);
 209                        if (*endptr) {
 210                                p_err("can't parse %s as CPU ID", **argv);
 211                                goto err_close_map;
 212                        }
 213
 214                        NEXT_ARG();
 215                } else if (is_prefix(*argv, "index")) {
 216                        char *endptr;
 217
 218                        NEXT_ARG();
 219                        index = strtoul(*argv, &endptr, 0);
 220                        if (*endptr) {
 221                                p_err("can't parse %s as index", **argv);
 222                                goto err_close_map;
 223                        }
 224
 225                        NEXT_ARG();
 226                } else {
 227                        BAD_ARG();
 228                        goto err_close_map;
 229                }
 230
 231                do_all = false;
 232        }
 233
 234        if (!do_all) {
 235                if (index == -1 || cpu == -1) {
 236                        p_err("cpu and index must be specified together");
 237                        goto err_close_map;
 238                }
 239
 240                nfds = 1;
 241        } else {
 242                nfds = min(get_possible_cpus(), map_info.max_entries);
 243                cpu = 0;
 244                index = 0;
 245        }
 246
 247        rings = calloc(nfds, sizeof(rings[0]));
 248        if (!rings)
 249                goto err_close_map;
 250
 251        pfds = calloc(nfds, sizeof(pfds[0]));
 252        if (!pfds)
 253                goto err_free_rings;
 254
 255        for (i = 0; i < nfds; i++) {
 256                rings[i].cpu = cpu + i;
 257                rings[i].key = index + i;
 258
 259                rings[i].fd = bpf_perf_event_open(map_fd, rings[i].key,
 260                                                  rings[i].cpu);
 261                if (rings[i].fd < 0)
 262                        goto err_close_fds_prev;
 263
 264                rings[i].mem = perf_event_mmap(rings[i].fd);
 265                if (!rings[i].mem)
 266                        goto err_close_fds_current;
 267
 268                pfds[i].fd = rings[i].fd;
 269                pfds[i].events = POLLIN;
 270        }
 271
 272        signal(SIGINT, int_exit);
 273        signal(SIGHUP, int_exit);
 274        signal(SIGTERM, int_exit);
 275
 276        if (json_output)
 277                jsonw_start_array(json_wtr);
 278
 279        while (!stop) {
 280                poll(pfds, nfds, 200);
 281                for (i = 0; i < nfds; i++)
 282                        perf_event_read(&rings[i], &tmp_buf, &tmp_buf_sz);
 283        }
 284        free(tmp_buf);
 285
 286        if (json_output)
 287                jsonw_end_array(json_wtr);
 288
 289        for (i = 0; i < nfds; i++) {
 290                perf_event_unmap(rings[i].mem);
 291                close(rings[i].fd);
 292        }
 293        free(pfds);
 294        free(rings);
 295        close(map_fd);
 296
 297        return 0;
 298
 299err_close_fds_prev:
 300        while (i--) {
 301                perf_event_unmap(rings[i].mem);
 302err_close_fds_current:
 303                close(rings[i].fd);
 304        }
 305        free(pfds);
 306err_free_rings:
 307        free(rings);
 308err_close_map:
 309        close(map_fd);
 310        return -1;
 311}
 312