^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1) // SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2) //
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3) // This file is provided under a dual BSD/GPLv2 license. When using or
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4) // redistributing this file, you may do so under either license.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 5) //
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 6) // Copyright(c) 2018 Intel Corporation. All rights reserved.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 7) //
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 8) // Author: Liam Girdwood <liam.r.girdwood@linux.intel.com>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 9) //
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 10)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 11) #include <linux/debugfs.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 12) #include <linux/sched/signal.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 13) #include "sof-priv.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 14) #include "ops.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 15)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 16) static size_t sof_trace_avail(struct snd_sof_dev *sdev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 17) loff_t pos, size_t buffer_size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 18) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 19) loff_t host_offset = READ_ONCE(sdev->host_offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 20)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 21) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 22) * If host offset is less than local pos, it means write pointer of
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 23) * host DMA buffer has been wrapped. We should output the trace data
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 24) * at the end of host DMA buffer at first.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 25) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 26) if (host_offset < pos)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 27) return buffer_size - pos;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 28)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 29) /* If there is available trace data now, it is unnecessary to wait. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 30) if (host_offset > pos)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 31) return host_offset - pos;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 32)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 33) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 34) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 35)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 36) static size_t sof_wait_trace_avail(struct snd_sof_dev *sdev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 37) loff_t pos, size_t buffer_size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 38) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 39) wait_queue_entry_t wait;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 40) size_t ret = sof_trace_avail(sdev, pos, buffer_size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 41)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 42) /* data immediately available */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 43) if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 44) return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 45)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 46) if (!sdev->dtrace_is_enabled && sdev->dtrace_draining) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 47) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 48) * tracing has ended and all traces have been
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 49) * read by client, return EOF
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 50) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 51) sdev->dtrace_draining = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 52) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 53) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 54)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 55) /* wait for available trace data from FW */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 56) init_waitqueue_entry(&wait, current);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 57) set_current_state(TASK_INTERRUPTIBLE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 58) add_wait_queue(&sdev->trace_sleep, &wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 59)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 60) if (!signal_pending(current)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 61) /* set timeout to max value, no error code */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 62) schedule_timeout(MAX_SCHEDULE_TIMEOUT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 63) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 64) remove_wait_queue(&sdev->trace_sleep, &wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 65)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 66) return sof_trace_avail(sdev, pos, buffer_size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 67) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 68)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 69) static ssize_t sof_dfsentry_trace_read(struct file *file, char __user *buffer,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 70) size_t count, loff_t *ppos)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 71) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 72) struct snd_sof_dfsentry *dfse = file->private_data;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 73) struct snd_sof_dev *sdev = dfse->sdev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 74) unsigned long rem;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 75) loff_t lpos = *ppos;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 76) size_t avail, buffer_size = dfse->size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 77) u64 lpos_64;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 78)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 79) /* make sure we know about any failures on the DSP side */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 80) sdev->dtrace_error = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 81)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 82) /* check pos and count */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 83) if (lpos < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 84) return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 85) if (!count)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 86) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 87)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 88) /* check for buffer wrap and count overflow */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 89) lpos_64 = lpos;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 90) lpos = do_div(lpos_64, buffer_size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 91)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 92) if (count > buffer_size - lpos) /* min() not used to avoid sparse warnings */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 93) count = buffer_size - lpos;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 94)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 95) /* get available count based on current host offset */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 96) avail = sof_wait_trace_avail(sdev, lpos, buffer_size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 97) if (sdev->dtrace_error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 98) dev_err(sdev->dev, "error: trace IO error\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 99) return -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102) /* make sure count is <= avail */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103) count = avail > count ? count : avail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) /* copy available trace data to debugfs */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106) rem = copy_to_user(buffer, ((u8 *)(dfse->buf) + lpos), count);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107) if (rem)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) return -EFAULT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) *ppos += count;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112) /* move debugfs reading position */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113) return count;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116) static int sof_dfsentry_trace_release(struct inode *inode, struct file *file)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) struct snd_sof_dfsentry *dfse = inode->i_private;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119) struct snd_sof_dev *sdev = dfse->sdev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121) /* avoid duplicate traces at next open */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122) if (!sdev->dtrace_is_enabled)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) sdev->host_offset = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128) static const struct file_operations sof_dfs_trace_fops = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129) .open = simple_open,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) .read = sof_dfsentry_trace_read,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131) .llseek = default_llseek,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132) .release = sof_dfsentry_trace_release,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135) static int trace_debugfs_create(struct snd_sof_dev *sdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137) struct snd_sof_dfsentry *dfse;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139) if (!sdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140) return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142) dfse = devm_kzalloc(sdev->dev, sizeof(*dfse), GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143) if (!dfse)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146) dfse->type = SOF_DFSENTRY_TYPE_BUF;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147) dfse->buf = sdev->dmatb.area;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148) dfse->size = sdev->dmatb.bytes;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149) dfse->sdev = sdev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151) debugfs_create_file("trace", 0444, sdev->debugfs_root, dfse,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) &sof_dfs_trace_fops);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157) int snd_sof_init_trace_ipc(struct snd_sof_dev *sdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159) struct sof_ipc_fw_ready *ready = &sdev->fw_ready;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) struct sof_ipc_fw_version *v = &ready->version;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161) struct sof_ipc_dma_trace_params_ext params;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162) struct sof_ipc_reply ipc_reply;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163) int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165) if (!sdev->dtrace_is_supported)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168) if (sdev->dtrace_is_enabled || !sdev->dma_trace_pages)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169) return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171) /* set IPC parameters */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172) params.hdr.cmd = SOF_IPC_GLB_TRACE_MSG;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173) /* PARAMS_EXT is only supported from ABI 3.7.0 onwards */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174) if (v->abi_version >= SOF_ABI_VER(3, 7, 0)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175) params.hdr.size = sizeof(struct sof_ipc_dma_trace_params_ext);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176) params.hdr.cmd |= SOF_IPC_TRACE_DMA_PARAMS_EXT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177) params.timestamp_ns = ktime_get(); /* in nanosecond */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178) } else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179) params.hdr.size = sizeof(struct sof_ipc_dma_trace_params);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180) params.hdr.cmd |= SOF_IPC_TRACE_DMA_PARAMS;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182) params.buffer.phy_addr = sdev->dmatp.addr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183) params.buffer.size = sdev->dmatb.bytes;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184) params.buffer.pages = sdev->dma_trace_pages;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185) params.stream_tag = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187) sdev->host_offset = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 188) sdev->dtrace_draining = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 189)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 190) ret = snd_sof_dma_trace_init(sdev, ¶ms.stream_tag);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 191) if (ret < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 192) dev_err(sdev->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 193) "error: fail in snd_sof_dma_trace_init %d\n", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 194) return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 195) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 196) dev_dbg(sdev->dev, "stream_tag: %d\n", params.stream_tag);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 197)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 198) /* send IPC to the DSP */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 199) ret = sof_ipc_tx_message(sdev->ipc,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 200) params.hdr.cmd, ¶ms, sizeof(params),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 201) &ipc_reply, sizeof(ipc_reply));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 202) if (ret < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 203) dev_err(sdev->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 204) "error: can't set params for DMA for trace %d\n", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 205) goto trace_release;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 206) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 207)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 208) ret = snd_sof_dma_trace_trigger(sdev, SNDRV_PCM_TRIGGER_START);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 209) if (ret < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 210) dev_err(sdev->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 211) "error: snd_sof_dma_trace_trigger: start: %d\n", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 212) goto trace_release;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 213) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 214)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 215) sdev->dtrace_is_enabled = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 216)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 217) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 218)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 219) trace_release:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 220) snd_sof_dma_trace_release(sdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 221) return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 222) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 223)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 224) int snd_sof_init_trace(struct snd_sof_dev *sdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 225) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 226) int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 227)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 228) if (!sdev->dtrace_is_supported)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 229) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 230)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 231) /* set false before start initialization */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 232) sdev->dtrace_is_enabled = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 233)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 234) /* allocate trace page table buffer */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 235) ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV, sdev->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 236) PAGE_SIZE, &sdev->dmatp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 237) if (ret < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 238) dev_err(sdev->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 239) "error: can't alloc page table for trace %d\n", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 240) return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 241) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 242)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 243) /* allocate trace data buffer */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 244) ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV_SG, sdev->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 245) DMA_BUF_SIZE_FOR_TRACE, &sdev->dmatb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 246) if (ret < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 247) dev_err(sdev->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 248) "error: can't alloc buffer for trace %d\n", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 249) goto page_err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 250) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 251)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 252) /* create compressed page table for audio firmware */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 253) ret = snd_sof_create_page_table(sdev->dev, &sdev->dmatb,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 254) sdev->dmatp.area, sdev->dmatb.bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 255) if (ret < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 256) goto table_err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 257)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 258) sdev->dma_trace_pages = ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 259) dev_dbg(sdev->dev, "dma_trace_pages: %d\n", sdev->dma_trace_pages);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 260)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 261) if (sdev->first_boot) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 262) ret = trace_debugfs_create(sdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 263) if (ret < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 264) goto table_err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 265) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 266)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 267) init_waitqueue_head(&sdev->trace_sleep);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 268)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 269) ret = snd_sof_init_trace_ipc(sdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 270) if (ret < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 271) goto table_err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 272)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 273) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 274) table_err:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 275) sdev->dma_trace_pages = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 276) snd_dma_free_pages(&sdev->dmatb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 277) page_err:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 278) snd_dma_free_pages(&sdev->dmatp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 279) return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 280) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 281) EXPORT_SYMBOL(snd_sof_init_trace);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 282)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 283) int snd_sof_trace_update_pos(struct snd_sof_dev *sdev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 284) struct sof_ipc_dma_trace_posn *posn)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 285) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 286) if (!sdev->dtrace_is_supported)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 287) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 288)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 289) if (sdev->dtrace_is_enabled && sdev->host_offset != posn->host_offset) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 290) sdev->host_offset = posn->host_offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 291) wake_up(&sdev->trace_sleep);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 292) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 293)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 294) if (posn->overflow != 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 295) dev_err(sdev->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 296) "error: DSP trace buffer overflow %u bytes. Total messages %d\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 297) posn->overflow, posn->messages);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 298)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 299) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 300) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 301)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 302) /* an error has occurred within the DSP that prevents further trace */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 303) void snd_sof_trace_notify_for_error(struct snd_sof_dev *sdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 304) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 305) if (!sdev->dtrace_is_supported)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 306) return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 307)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 308) if (sdev->dtrace_is_enabled) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 309) dev_err(sdev->dev, "error: waking up any trace sleepers\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 310) sdev->dtrace_error = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 311) wake_up(&sdev->trace_sleep);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 312) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 313) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 314) EXPORT_SYMBOL(snd_sof_trace_notify_for_error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 315)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 316) void snd_sof_release_trace(struct snd_sof_dev *sdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 317) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 318) int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 319)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 320) if (!sdev->dtrace_is_supported || !sdev->dtrace_is_enabled)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 321) return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 322)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 323) ret = snd_sof_dma_trace_trigger(sdev, SNDRV_PCM_TRIGGER_STOP);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 324) if (ret < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 325) dev_err(sdev->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 326) "error: snd_sof_dma_trace_trigger: stop: %d\n", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 327)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 328) ret = snd_sof_dma_trace_release(sdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 329) if (ret < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 330) dev_err(sdev->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 331) "error: fail in snd_sof_dma_trace_release %d\n", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 332)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 333) sdev->dtrace_is_enabled = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 334) sdev->dtrace_draining = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 335) wake_up(&sdev->trace_sleep);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 336) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 337) EXPORT_SYMBOL(snd_sof_release_trace);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 338)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 339) void snd_sof_free_trace(struct snd_sof_dev *sdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 340) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 341) if (!sdev->dtrace_is_supported)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 342) return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 343)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 344) snd_sof_release_trace(sdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 345)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 346) if (sdev->dma_trace_pages) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 347) snd_dma_free_pages(&sdev->dmatb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 348) snd_dma_free_pages(&sdev->dmatp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 349) sdev->dma_trace_pages = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 350) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 351) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 352) EXPORT_SYMBOL(snd_sof_free_trace);