linux/sound/soc/sof/trace.c
<<
>>
Prefs
   1// SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause)
   2//
   3// This file is provided under a dual BSD/GPLv2 license.  When using or
   4// redistributing this file, you may do so under either license.
   5//
   6// Copyright(c) 2018 Intel Corporation. All rights reserved.
   7//
   8// Author: Liam Girdwood <liam.r.girdwood@linux.intel.com>
   9//
  10
  11#include <linux/debugfs.h>
  12#include <linux/sched/signal.h>
  13#include "sof-priv.h"
  14#include "ops.h"
  15
  16#define TRACE_FILTER_ELEMENTS_PER_ENTRY 4
  17#define TRACE_FILTER_MAX_CONFIG_STRING_LENGTH 1024
  18
  19static int trace_filter_append_elem(struct snd_sof_dev *sdev, uint32_t key, uint32_t value,
  20                                    struct sof_ipc_trace_filter_elem *elem_list,
  21                                    int capacity, int *counter)
  22{
  23        if (*counter >= capacity)
  24                return -ENOMEM;
  25
  26        elem_list[*counter].key = key;
  27        elem_list[*counter].value = value;
  28        ++*counter;
  29
  30        return 0;
  31}
  32
  33static int trace_filter_parse_entry(struct snd_sof_dev *sdev, const char *line,
  34                                    struct sof_ipc_trace_filter_elem *elem,
  35                                    int capacity, int *counter)
  36{
  37        int len = strlen(line);
  38        int cnt = *counter;
  39        uint32_t uuid_id;
  40        int log_level;
  41        int pipe_id;
  42        int comp_id;
  43        int read;
  44        int ret;
  45
  46        /* ignore empty content */
  47        ret = sscanf(line, " %n", &read);
  48        if (!ret && read == len)
  49                return len;
  50
  51        ret = sscanf(line, " %d %x %d %d %n", &log_level, &uuid_id, &pipe_id, &comp_id, &read);
  52        if (ret != TRACE_FILTER_ELEMENTS_PER_ENTRY || read != len) {
  53                dev_err(sdev->dev, "error: invalid trace filter entry '%s'\n", line);
  54                return -EINVAL;
  55        }
  56
  57        if (uuid_id > 0) {
  58                ret = trace_filter_append_elem(sdev, SOF_IPC_TRACE_FILTER_ELEM_BY_UUID,
  59                                               uuid_id, elem, capacity, &cnt);
  60                if (ret)
  61                        return ret;
  62        }
  63        if (pipe_id >= 0) {
  64                ret = trace_filter_append_elem(sdev, SOF_IPC_TRACE_FILTER_ELEM_BY_PIPE,
  65                                               pipe_id, elem, capacity, &cnt);
  66                if (ret)
  67                        return ret;
  68        }
  69        if (comp_id >= 0) {
  70                ret = trace_filter_append_elem(sdev, SOF_IPC_TRACE_FILTER_ELEM_BY_COMP,
  71                                               comp_id, elem, capacity, &cnt);
  72                if (ret)
  73                        return ret;
  74        }
  75
  76        ret = trace_filter_append_elem(sdev, SOF_IPC_TRACE_FILTER_ELEM_SET_LEVEL |
  77                                       SOF_IPC_TRACE_FILTER_ELEM_FIN,
  78                                       log_level, elem, capacity, &cnt);
  79        if (ret)
  80                return ret;
  81
  82        /* update counter only when parsing whole entry passed */
  83        *counter = cnt;
  84
  85        return len;
  86}
  87
  88static int trace_filter_parse(struct snd_sof_dev *sdev, char *string,
  89                              int *out_elem_cnt,
  90                              struct sof_ipc_trace_filter_elem **out)
  91{
  92        static const char entry_delimiter[] = ";";
  93        char *entry = string;
  94        int capacity = 0;
  95        int entry_len;
  96        int cnt = 0;
  97
  98        /*
  99         * Each entry contains at least 1, up to TRACE_FILTER_ELEMENTS_PER_ENTRY
 100         * IPC elements, depending on content. Calculate IPC elements capacity
 101         * for the input string where each element is set.
 102         */
 103        while (entry) {
 104                capacity += TRACE_FILTER_ELEMENTS_PER_ENTRY;
 105                entry = strchr(entry + 1, entry_delimiter[0]);
 106        }
 107        *out = kmalloc(capacity * sizeof(**out), GFP_KERNEL);
 108        if (!*out)
 109                return -ENOMEM;
 110
 111        /* split input string by ';', and parse each entry separately in trace_filter_parse_entry */
 112        while ((entry = strsep(&string, entry_delimiter))) {
 113                entry_len = trace_filter_parse_entry(sdev, entry, *out, capacity, &cnt);
 114                if (entry_len < 0) {
 115                        dev_err(sdev->dev, "error: %s failed for '%s', %d\n", __func__, entry,
 116                                entry_len);
 117                        return -EINVAL;
 118                }
 119        }
 120
 121        *out_elem_cnt = cnt;
 122
 123        return 0;
 124}
 125
 126static int sof_ipc_trace_update_filter(struct snd_sof_dev *sdev, int num_elems,
 127                                       struct sof_ipc_trace_filter_elem *elems)
 128{
 129        struct sof_ipc_trace_filter *msg;
 130        struct sof_ipc_reply reply;
 131        size_t size;
 132        int ret;
 133
 134        size = struct_size(msg, elems, num_elems);
 135        if (size > SOF_IPC_MSG_MAX_SIZE)
 136                return -EINVAL;
 137
 138        msg = kmalloc(size, GFP_KERNEL);
 139        if (!msg)
 140                return -ENOMEM;
 141
 142        msg->hdr.size = size;
 143        msg->hdr.cmd = SOF_IPC_GLB_TRACE_MSG | SOF_IPC_TRACE_FILTER_UPDATE;
 144        msg->elem_cnt = num_elems;
 145        memcpy(&msg->elems[0], elems, num_elems * sizeof(*elems));
 146
 147        ret = pm_runtime_get_sync(sdev->dev);
 148        if (ret < 0 && ret != -EACCES) {
 149                pm_runtime_put_noidle(sdev->dev);
 150                dev_err(sdev->dev, "error: enabling device failed: %d\n", ret);
 151                goto error;
 152        }
 153        ret = sof_ipc_tx_message(sdev->ipc, msg->hdr.cmd, msg, msg->hdr.size,
 154                                 &reply, sizeof(reply));
 155        pm_runtime_mark_last_busy(sdev->dev);
 156        pm_runtime_put_autosuspend(sdev->dev);
 157
 158error:
 159        kfree(msg);
 160        return ret ? ret : reply.error;
 161}
 162
 163static ssize_t sof_dfsentry_trace_filter_write(struct file *file, const char __user *from,
 164                                               size_t count, loff_t *ppos)
 165{
 166        struct snd_sof_dfsentry *dfse = file->private_data;
 167        struct sof_ipc_trace_filter_elem *elems = NULL;
 168        struct snd_sof_dev *sdev = dfse->sdev;
 169        loff_t pos = 0;
 170        int num_elems;
 171        char *string;
 172        int ret;
 173
 174        if (count > TRACE_FILTER_MAX_CONFIG_STRING_LENGTH) {
 175                dev_err(sdev->dev, "%s too long input, %zu > %d\n", __func__, count,
 176                        TRACE_FILTER_MAX_CONFIG_STRING_LENGTH);
 177                return -EINVAL;
 178        }
 179
 180        string = kmalloc(count + 1, GFP_KERNEL);
 181        if (!string)
 182                return -ENOMEM;
 183
 184        /* assert null termination */
 185        string[count] = 0;
 186        ret = simple_write_to_buffer(string, count, &pos, from, count);
 187        if (ret < 0)
 188                goto error;
 189
 190        ret = trace_filter_parse(sdev, string, &num_elems, &elems);
 191        if (ret < 0) {
 192                dev_err(sdev->dev, "error: fail in trace_filter_parse, %d\n", ret);
 193                goto error;
 194        }
 195
 196        if (num_elems) {
 197                ret = sof_ipc_trace_update_filter(sdev, num_elems, elems);
 198                if (ret < 0) {
 199                        dev_err(sdev->dev, "error: fail in sof_ipc_trace_update_filter %d\n", ret);
 200                        goto error;
 201                }
 202        }
 203        ret = count;
 204error:
 205        kfree(string);
 206        kfree(elems);
 207        return ret;
 208}
 209
 210static const struct file_operations sof_dfs_trace_filter_fops = {
 211        .open = simple_open,
 212        .write = sof_dfsentry_trace_filter_write,
 213        .llseek = default_llseek,
 214};
 215
 216static int trace_debugfs_filter_create(struct snd_sof_dev *sdev)
 217{
 218        struct snd_sof_dfsentry *dfse;
 219
 220        dfse = devm_kzalloc(sdev->dev, sizeof(*dfse), GFP_KERNEL);
 221        if (!dfse)
 222                return -ENOMEM;
 223
 224        dfse->sdev = sdev;
 225        dfse->type = SOF_DFSENTRY_TYPE_BUF;
 226
 227        debugfs_create_file("filter", 0200, sdev->debugfs_root, dfse,
 228                            &sof_dfs_trace_filter_fops);
 229        /* add to dfsentry list */
 230        list_add(&dfse->list, &sdev->dfsentry_list);
 231
 232        return 0;
 233}
 234
 235static size_t sof_trace_avail(struct snd_sof_dev *sdev,
 236                              loff_t pos, size_t buffer_size)
 237{
 238        loff_t host_offset = READ_ONCE(sdev->host_offset);
 239
 240        /*
 241         * If host offset is less than local pos, it means write pointer of
 242         * host DMA buffer has been wrapped. We should output the trace data
 243         * at the end of host DMA buffer at first.
 244         */
 245        if (host_offset < pos)
 246                return buffer_size - pos;
 247
 248        /* If there is available trace data now, it is unnecessary to wait. */
 249        if (host_offset > pos)
 250                return host_offset - pos;
 251
 252        return 0;
 253}
 254
 255static size_t sof_wait_trace_avail(struct snd_sof_dev *sdev,
 256                                   loff_t pos, size_t buffer_size)
 257{
 258        wait_queue_entry_t wait;
 259        size_t ret = sof_trace_avail(sdev, pos, buffer_size);
 260
 261        /* data immediately available */
 262        if (ret)
 263                return ret;
 264
 265        if (!sdev->dtrace_is_enabled && sdev->dtrace_draining) {
 266                /*
 267                 * tracing has ended and all traces have been
 268                 * read by client, return EOF
 269                 */
 270                sdev->dtrace_draining = false;
 271                return 0;
 272        }
 273
 274        /* wait for available trace data from FW */
 275        init_waitqueue_entry(&wait, current);
 276        set_current_state(TASK_INTERRUPTIBLE);
 277        add_wait_queue(&sdev->trace_sleep, &wait);
 278
 279        if (!signal_pending(current)) {
 280                /* set timeout to max value, no error code */
 281                schedule_timeout(MAX_SCHEDULE_TIMEOUT);
 282        }
 283        remove_wait_queue(&sdev->trace_sleep, &wait);
 284
 285        return sof_trace_avail(sdev, pos, buffer_size);
 286}
 287
 288static ssize_t sof_dfsentry_trace_read(struct file *file, char __user *buffer,
 289                                       size_t count, loff_t *ppos)
 290{
 291        struct snd_sof_dfsentry *dfse = file->private_data;
 292        struct snd_sof_dev *sdev = dfse->sdev;
 293        unsigned long rem;
 294        loff_t lpos = *ppos;
 295        size_t avail, buffer_size = dfse->size;
 296        u64 lpos_64;
 297
 298        /* make sure we know about any failures on the DSP side */
 299        sdev->dtrace_error = false;
 300
 301        /* check pos and count */
 302        if (lpos < 0)
 303                return -EINVAL;
 304        if (!count)
 305                return 0;
 306
 307        /* check for buffer wrap and count overflow */
 308        lpos_64 = lpos;
 309        lpos = do_div(lpos_64, buffer_size);
 310
 311        if (count > buffer_size - lpos) /* min() not used to avoid sparse warnings */
 312                count = buffer_size - lpos;
 313
 314        /* get available count based on current host offset */
 315        avail = sof_wait_trace_avail(sdev, lpos, buffer_size);
 316        if (sdev->dtrace_error) {
 317                dev_err(sdev->dev, "error: trace IO error\n");
 318                return -EIO;
 319        }
 320
 321        /* make sure count is <= avail */
 322        count = avail > count ? count : avail;
 323
 324        /* copy available trace data to debugfs */
 325        rem = copy_to_user(buffer, ((u8 *)(dfse->buf) + lpos), count);
 326        if (rem)
 327                return -EFAULT;
 328
 329        *ppos += count;
 330
 331        /* move debugfs reading position */
 332        return count;
 333}
 334
 335static int sof_dfsentry_trace_release(struct inode *inode, struct file *file)
 336{
 337        struct snd_sof_dfsentry *dfse = inode->i_private;
 338        struct snd_sof_dev *sdev = dfse->sdev;
 339
 340        /* avoid duplicate traces at next open */
 341        if (!sdev->dtrace_is_enabled)
 342                sdev->host_offset = 0;
 343
 344        return 0;
 345}
 346
 347static const struct file_operations sof_dfs_trace_fops = {
 348        .open = simple_open,
 349        .read = sof_dfsentry_trace_read,
 350        .llseek = default_llseek,
 351        .release = sof_dfsentry_trace_release,
 352};
 353
 354static int trace_debugfs_create(struct snd_sof_dev *sdev)
 355{
 356        struct snd_sof_dfsentry *dfse;
 357        int ret;
 358
 359        if (!sdev)
 360                return -EINVAL;
 361
 362        ret = trace_debugfs_filter_create(sdev);
 363        if (ret < 0)
 364                dev_err(sdev->dev, "error: fail in %s, %d", __func__, ret);
 365
 366        dfse = devm_kzalloc(sdev->dev, sizeof(*dfse), GFP_KERNEL);
 367        if (!dfse)
 368                return -ENOMEM;
 369
 370        dfse->type = SOF_DFSENTRY_TYPE_BUF;
 371        dfse->buf = sdev->dmatb.area;
 372        dfse->size = sdev->dmatb.bytes;
 373        dfse->sdev = sdev;
 374
 375        debugfs_create_file("trace", 0444, sdev->debugfs_root, dfse,
 376                            &sof_dfs_trace_fops);
 377
 378        return 0;
 379}
 380
 381int snd_sof_init_trace_ipc(struct snd_sof_dev *sdev)
 382{
 383        struct sof_ipc_fw_ready *ready = &sdev->fw_ready;
 384        struct sof_ipc_fw_version *v = &ready->version;
 385        struct sof_ipc_dma_trace_params_ext params;
 386        struct sof_ipc_reply ipc_reply;
 387        int ret;
 388
 389        if (!sdev->dtrace_is_supported)
 390                return 0;
 391
 392        if (sdev->dtrace_is_enabled || !sdev->dma_trace_pages)
 393                return -EINVAL;
 394
 395        /* set IPC parameters */
 396        params.hdr.cmd = SOF_IPC_GLB_TRACE_MSG;
 397        /* PARAMS_EXT is only supported from ABI 3.7.0 onwards */
 398        if (v->abi_version >= SOF_ABI_VER(3, 7, 0)) {
 399                params.hdr.size = sizeof(struct sof_ipc_dma_trace_params_ext);
 400                params.hdr.cmd |= SOF_IPC_TRACE_DMA_PARAMS_EXT;
 401                params.timestamp_ns = ktime_get(); /* in nanosecond */
 402        } else {
 403                params.hdr.size = sizeof(struct sof_ipc_dma_trace_params);
 404                params.hdr.cmd |= SOF_IPC_TRACE_DMA_PARAMS;
 405        }
 406        params.buffer.phy_addr = sdev->dmatp.addr;
 407        params.buffer.size = sdev->dmatb.bytes;
 408        params.buffer.pages = sdev->dma_trace_pages;
 409        params.stream_tag = 0;
 410
 411        sdev->host_offset = 0;
 412        sdev->dtrace_draining = false;
 413
 414        ret = snd_sof_dma_trace_init(sdev, &params.stream_tag);
 415        if (ret < 0) {
 416                dev_err(sdev->dev,
 417                        "error: fail in snd_sof_dma_trace_init %d\n", ret);
 418                return ret;
 419        }
 420        dev_dbg(sdev->dev, "%s: stream_tag: %d\n", __func__, params.stream_tag);
 421
 422        /* send IPC to the DSP */
 423        ret = sof_ipc_tx_message(sdev->ipc,
 424                                 params.hdr.cmd, &params, sizeof(params),
 425                                 &ipc_reply, sizeof(ipc_reply));
 426        if (ret < 0) {
 427                dev_err(sdev->dev,
 428                        "error: can't set params for DMA for trace %d\n", ret);
 429                goto trace_release;
 430        }
 431
 432        ret = snd_sof_dma_trace_trigger(sdev, SNDRV_PCM_TRIGGER_START);
 433        if (ret < 0) {
 434                dev_err(sdev->dev,
 435                        "error: snd_sof_dma_trace_trigger: start: %d\n", ret);
 436                goto trace_release;
 437        }
 438
 439        sdev->dtrace_is_enabled = true;
 440
 441        return 0;
 442
 443trace_release:
 444        snd_sof_dma_trace_release(sdev);
 445        return ret;
 446}
 447
 448int snd_sof_init_trace(struct snd_sof_dev *sdev)
 449{
 450        int ret;
 451
 452        if (!sdev->dtrace_is_supported)
 453                return 0;
 454
 455        /* set false before start initialization */
 456        sdev->dtrace_is_enabled = false;
 457
 458        /* allocate trace page table buffer */
 459        ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV, sdev->dev,
 460                                  PAGE_SIZE, &sdev->dmatp);
 461        if (ret < 0) {
 462                dev_err(sdev->dev,
 463                        "error: can't alloc page table for trace %d\n", ret);
 464                return ret;
 465        }
 466
 467        /* allocate trace data buffer */
 468        ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV_SG, sdev->dev,
 469                                  DMA_BUF_SIZE_FOR_TRACE, &sdev->dmatb);
 470        if (ret < 0) {
 471                dev_err(sdev->dev,
 472                        "error: can't alloc buffer for trace %d\n", ret);
 473                goto page_err;
 474        }
 475
 476        /* create compressed page table for audio firmware */
 477        ret = snd_sof_create_page_table(sdev->dev, &sdev->dmatb,
 478                                        sdev->dmatp.area, sdev->dmatb.bytes);
 479        if (ret < 0)
 480                goto table_err;
 481
 482        sdev->dma_trace_pages = ret;
 483        dev_dbg(sdev->dev, "%s: dma_trace_pages: %d\n",
 484                __func__, sdev->dma_trace_pages);
 485
 486        if (sdev->first_boot) {
 487                ret = trace_debugfs_create(sdev);
 488                if (ret < 0)
 489                        goto table_err;
 490        }
 491
 492        init_waitqueue_head(&sdev->trace_sleep);
 493
 494        ret = snd_sof_init_trace_ipc(sdev);
 495        if (ret < 0)
 496                goto table_err;
 497
 498        return 0;
 499table_err:
 500        sdev->dma_trace_pages = 0;
 501        snd_dma_free_pages(&sdev->dmatb);
 502page_err:
 503        snd_dma_free_pages(&sdev->dmatp);
 504        return ret;
 505}
 506EXPORT_SYMBOL(snd_sof_init_trace);
 507
 508int snd_sof_trace_update_pos(struct snd_sof_dev *sdev,
 509                             struct sof_ipc_dma_trace_posn *posn)
 510{
 511        if (!sdev->dtrace_is_supported)
 512                return 0;
 513
 514        if (sdev->dtrace_is_enabled && sdev->host_offset != posn->host_offset) {
 515                sdev->host_offset = posn->host_offset;
 516                wake_up(&sdev->trace_sleep);
 517        }
 518
 519        if (posn->overflow != 0)
 520                dev_err(sdev->dev,
 521                        "error: DSP trace buffer overflow %u bytes. Total messages %d\n",
 522                        posn->overflow, posn->messages);
 523
 524        return 0;
 525}
 526
 527/* an error has occurred within the DSP that prevents further trace */
 528void snd_sof_trace_notify_for_error(struct snd_sof_dev *sdev)
 529{
 530        if (!sdev->dtrace_is_supported)
 531                return;
 532
 533        if (sdev->dtrace_is_enabled) {
 534                sdev->dtrace_error = true;
 535                wake_up(&sdev->trace_sleep);
 536        }
 537}
 538EXPORT_SYMBOL(snd_sof_trace_notify_for_error);
 539
 540void snd_sof_release_trace(struct snd_sof_dev *sdev)
 541{
 542        int ret;
 543
 544        if (!sdev->dtrace_is_supported || !sdev->dtrace_is_enabled)
 545                return;
 546
 547        ret = snd_sof_dma_trace_trigger(sdev, SNDRV_PCM_TRIGGER_STOP);
 548        if (ret < 0)
 549                dev_err(sdev->dev,
 550                        "error: snd_sof_dma_trace_trigger: stop: %d\n", ret);
 551
 552        ret = snd_sof_dma_trace_release(sdev);
 553        if (ret < 0)
 554                dev_err(sdev->dev,
 555                        "error: fail in snd_sof_dma_trace_release %d\n", ret);
 556
 557        sdev->dtrace_is_enabled = false;
 558        sdev->dtrace_draining = true;
 559        wake_up(&sdev->trace_sleep);
 560}
 561EXPORT_SYMBOL(snd_sof_release_trace);
 562
 563void snd_sof_free_trace(struct snd_sof_dev *sdev)
 564{
 565        if (!sdev->dtrace_is_supported)
 566                return;
 567
 568        snd_sof_release_trace(sdev);
 569
 570        if (sdev->dma_trace_pages) {
 571                snd_dma_free_pages(&sdev->dmatb);
 572                snd_dma_free_pages(&sdev->dmatp);
 573                sdev->dma_trace_pages = 0;
 574        }
 575}
 576EXPORT_SYMBOL(snd_sof_free_trace);
 577