linux/tools/testing/selftests/bpf/prog_tests/map_init.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0-only
   2/* Copyright (c) 2020 Tessares SA <http://www.tessares.net> */
   3
   4#include <test_progs.h>
   5#include "test_map_init.skel.h"
   6
   7#define TEST_VALUE 0x1234
   8#define FILL_VALUE 0xdeadbeef
   9
  10static int nr_cpus;
  11static int duration;
  12
  13typedef unsigned long long map_key_t;
  14typedef unsigned long long map_value_t;
  15typedef struct {
  16        map_value_t v; /* padding */
  17} __bpf_percpu_val_align pcpu_map_value_t;
  18
  19
  20static int map_populate(int map_fd, int num)
  21{
  22        pcpu_map_value_t value[nr_cpus];
  23        int i, err;
  24        map_key_t key;
  25
  26        for (i = 0; i < nr_cpus; i++)
  27                bpf_percpu(value, i) = FILL_VALUE;
  28
  29        for (key = 1; key <= num; key++) {
  30                err = bpf_map_update_elem(map_fd, &key, value, BPF_NOEXIST);
  31                if (!ASSERT_OK(err, "bpf_map_update_elem"))
  32                        return -1;
  33        }
  34
  35        return 0;
  36}
  37
  38static struct test_map_init *setup(enum bpf_map_type map_type, int map_sz,
  39                            int *map_fd, int populate)
  40{
  41        struct test_map_init *skel;
  42        int err;
  43
  44        skel = test_map_init__open();
  45        if (!ASSERT_OK_PTR(skel, "skel_open"))
  46                return NULL;
  47
  48        err = bpf_map__set_type(skel->maps.hashmap1, map_type);
  49        if (!ASSERT_OK(err, "bpf_map__set_type"))
  50                goto error;
  51
  52        err = bpf_map__set_max_entries(skel->maps.hashmap1, map_sz);
  53        if (!ASSERT_OK(err, "bpf_map__set_max_entries"))
  54                goto error;
  55
  56        err = test_map_init__load(skel);
  57        if (!ASSERT_OK(err, "skel_load"))
  58                goto error;
  59
  60        *map_fd = bpf_map__fd(skel->maps.hashmap1);
  61        if (CHECK(*map_fd < 0, "bpf_map__fd", "failed\n"))
  62                goto error;
  63
  64        err = map_populate(*map_fd, populate);
  65        if (!ASSERT_OK(err, "map_populate"))
  66                goto error_map;
  67
  68        return skel;
  69
  70error_map:
  71        close(*map_fd);
  72error:
  73        test_map_init__destroy(skel);
  74        return NULL;
  75}
  76
  77/* executes bpf program that updates map with key, value */
  78static int prog_run_insert_elem(struct test_map_init *skel, map_key_t key,
  79                                map_value_t value)
  80{
  81        struct test_map_init__bss *bss;
  82
  83        bss = skel->bss;
  84
  85        bss->inKey = key;
  86        bss->inValue = value;
  87        bss->inPid = getpid();
  88
  89        if (!ASSERT_OK(test_map_init__attach(skel), "skel_attach"))
  90                return -1;
  91
  92        /* Let tracepoint trigger */
  93        syscall(__NR_getpgid);
  94
  95        test_map_init__detach(skel);
  96
  97        return 0;
  98}
  99
 100static int check_values_one_cpu(pcpu_map_value_t *value, map_value_t expected)
 101{
 102        int i, nzCnt = 0;
 103        map_value_t val;
 104
 105        for (i = 0; i < nr_cpus; i++) {
 106                val = bpf_percpu(value, i);
 107                if (val) {
 108                        if (CHECK(val != expected, "map value",
 109                                  "unexpected for cpu %d: 0x%llx\n", i, val))
 110                                return -1;
 111                        nzCnt++;
 112                }
 113        }
 114
 115        if (CHECK(nzCnt != 1, "map value", "set for %d CPUs instead of 1!\n",
 116                  nzCnt))
 117                return -1;
 118
 119        return 0;
 120}
 121
 122/* Add key=1 elem with values set for all CPUs
 123 * Delete elem key=1
 124 * Run bpf prog that inserts new key=1 elem with value=0x1234
 125 *   (bpf prog can only set value for current CPU)
 126 * Lookup Key=1 and check value is as expected for all CPUs:
 127 *   value set by bpf prog for one CPU, 0 for all others
 128 */
 129static void test_pcpu_map_init(void)
 130{
 131        pcpu_map_value_t value[nr_cpus];
 132        struct test_map_init *skel;
 133        int map_fd, err;
 134        map_key_t key;
 135
 136        /* max 1 elem in map so insertion is forced to reuse freed entry */
 137        skel = setup(BPF_MAP_TYPE_PERCPU_HASH, 1, &map_fd, 1);
 138        if (!ASSERT_OK_PTR(skel, "prog_setup"))
 139                return;
 140
 141        /* delete element so the entry can be re-used*/
 142        key = 1;
 143        err = bpf_map_delete_elem(map_fd, &key);
 144        if (!ASSERT_OK(err, "bpf_map_delete_elem"))
 145                goto cleanup;
 146
 147        /* run bpf prog that inserts new elem, re-using the slot just freed */
 148        err = prog_run_insert_elem(skel, key, TEST_VALUE);
 149        if (!ASSERT_OK(err, "prog_run_insert_elem"))
 150                goto cleanup;
 151
 152        /* check that key=1 was re-created by bpf prog */
 153        err = bpf_map_lookup_elem(map_fd, &key, value);
 154        if (!ASSERT_OK(err, "bpf_map_lookup_elem"))
 155                goto cleanup;
 156
 157        /* and has expected values */
 158        check_values_one_cpu(value, TEST_VALUE);
 159
 160cleanup:
 161        test_map_init__destroy(skel);
 162}
 163
 164/* Add key=1 and key=2 elems with values set for all CPUs
 165 * Run bpf prog that inserts new key=3 elem
 166 *   (only for current cpu; other cpus should have initial value = 0)
 167 * Lookup Key=1 and check value is as expected for all CPUs
 168 */
 169static void test_pcpu_lru_map_init(void)
 170{
 171        pcpu_map_value_t value[nr_cpus];
 172        struct test_map_init *skel;
 173        int map_fd, err;
 174        map_key_t key;
 175
 176        /* Set up LRU map with 2 elements, values filled for all CPUs.
 177         * With these 2 elements, the LRU map is full
 178         */
 179        skel = setup(BPF_MAP_TYPE_LRU_PERCPU_HASH, 2, &map_fd, 2);
 180        if (!ASSERT_OK_PTR(skel, "prog_setup"))
 181                return;
 182
 183        /* run bpf prog that inserts new key=3 element, re-using LRU slot */
 184        key = 3;
 185        err = prog_run_insert_elem(skel, key, TEST_VALUE);
 186        if (!ASSERT_OK(err, "prog_run_insert_elem"))
 187                goto cleanup;
 188
 189        /* check that key=3 replaced one of earlier elements */
 190        err = bpf_map_lookup_elem(map_fd, &key, value);
 191        if (!ASSERT_OK(err, "bpf_map_lookup_elem"))
 192                goto cleanup;
 193
 194        /* and has expected values */
 195        check_values_one_cpu(value, TEST_VALUE);
 196
 197cleanup:
 198        test_map_init__destroy(skel);
 199}
 200
 201void test_map_init(void)
 202{
 203        nr_cpus = bpf_num_possible_cpus();
 204        if (nr_cpus <= 1) {
 205                printf("%s:SKIP: >1 cpu needed for this test\n", __func__);
 206                test__skip();
 207                return;
 208        }
 209
 210        if (test__start_subtest("pcpu_map_init"))
 211                test_pcpu_map_init();
 212        if (test__start_subtest("pcpu_lru_map_init"))
 213                test_pcpu_lru_map_init();
 214}
 215