David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame^] | 1 | // SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) |
| 2 | // |
| 3 | // This file is provided under a dual BSD/GPLv2 license. When using or |
| 4 | // redistributing this file, you may do so under either license. |
| 5 | // |
| 6 | // Copyright(c) 2018 Intel Corporation. All rights reserved. |
| 7 | // |
| 8 | // Author: Liam Girdwood <liam.r.girdwood@linux.intel.com> |
| 9 | // |
| 10 | |
| 11 | #include <linux/debugfs.h> |
| 12 | #include <linux/sched/signal.h> |
| 13 | #include "sof-priv.h" |
| 14 | #include "ops.h" |
| 15 | |
| 16 | static size_t sof_trace_avail(struct snd_sof_dev *sdev, |
| 17 | loff_t pos, size_t buffer_size) |
| 18 | { |
| 19 | loff_t host_offset = READ_ONCE(sdev->host_offset); |
| 20 | |
| 21 | /* |
| 22 | * If host offset is less than local pos, it means write pointer of |
| 23 | * host DMA buffer has been wrapped. We should output the trace data |
| 24 | * at the end of host DMA buffer at first. |
| 25 | */ |
| 26 | if (host_offset < pos) |
| 27 | return buffer_size - pos; |
| 28 | |
| 29 | /* If there is available trace data now, it is unnecessary to wait. */ |
| 30 | if (host_offset > pos) |
| 31 | return host_offset - pos; |
| 32 | |
| 33 | return 0; |
| 34 | } |
| 35 | |
| 36 | static size_t sof_wait_trace_avail(struct snd_sof_dev *sdev, |
| 37 | loff_t pos, size_t buffer_size) |
| 38 | { |
| 39 | wait_queue_entry_t wait; |
| 40 | size_t ret = sof_trace_avail(sdev, pos, buffer_size); |
| 41 | |
| 42 | /* data immediately available */ |
| 43 | if (ret) |
| 44 | return ret; |
| 45 | |
| 46 | if (!sdev->dtrace_is_enabled && sdev->dtrace_draining) { |
| 47 | /* |
| 48 | * tracing has ended and all traces have been |
| 49 | * read by client, return EOF |
| 50 | */ |
| 51 | sdev->dtrace_draining = false; |
| 52 | return 0; |
| 53 | } |
| 54 | |
| 55 | /* wait for available trace data from FW */ |
| 56 | init_waitqueue_entry(&wait, current); |
| 57 | set_current_state(TASK_INTERRUPTIBLE); |
| 58 | add_wait_queue(&sdev->trace_sleep, &wait); |
| 59 | |
| 60 | if (!signal_pending(current)) { |
| 61 | /* set timeout to max value, no error code */ |
| 62 | schedule_timeout(MAX_SCHEDULE_TIMEOUT); |
| 63 | } |
| 64 | remove_wait_queue(&sdev->trace_sleep, &wait); |
| 65 | |
| 66 | return sof_trace_avail(sdev, pos, buffer_size); |
| 67 | } |
| 68 | |
| 69 | static ssize_t sof_dfsentry_trace_read(struct file *file, char __user *buffer, |
| 70 | size_t count, loff_t *ppos) |
| 71 | { |
| 72 | struct snd_sof_dfsentry *dfse = file->private_data; |
| 73 | struct snd_sof_dev *sdev = dfse->sdev; |
| 74 | unsigned long rem; |
| 75 | loff_t lpos = *ppos; |
| 76 | size_t avail, buffer_size = dfse->size; |
| 77 | u64 lpos_64; |
| 78 | |
| 79 | /* make sure we know about any failures on the DSP side */ |
| 80 | sdev->dtrace_error = false; |
| 81 | |
| 82 | /* check pos and count */ |
| 83 | if (lpos < 0) |
| 84 | return -EINVAL; |
| 85 | if (!count) |
| 86 | return 0; |
| 87 | |
| 88 | /* check for buffer wrap and count overflow */ |
| 89 | lpos_64 = lpos; |
| 90 | lpos = do_div(lpos_64, buffer_size); |
| 91 | |
| 92 | if (count > buffer_size - lpos) /* min() not used to avoid sparse warnings */ |
| 93 | count = buffer_size - lpos; |
| 94 | |
| 95 | /* get available count based on current host offset */ |
| 96 | avail = sof_wait_trace_avail(sdev, lpos, buffer_size); |
| 97 | if (sdev->dtrace_error) { |
| 98 | dev_err(sdev->dev, "error: trace IO error\n"); |
| 99 | return -EIO; |
| 100 | } |
| 101 | |
| 102 | /* make sure count is <= avail */ |
| 103 | count = avail > count ? count : avail; |
| 104 | |
| 105 | /* copy available trace data to debugfs */ |
| 106 | rem = copy_to_user(buffer, ((u8 *)(dfse->buf) + lpos), count); |
| 107 | if (rem) |
| 108 | return -EFAULT; |
| 109 | |
| 110 | *ppos += count; |
| 111 | |
| 112 | /* move debugfs reading position */ |
| 113 | return count; |
| 114 | } |
| 115 | |
| 116 | static int sof_dfsentry_trace_release(struct inode *inode, struct file *file) |
| 117 | { |
| 118 | struct snd_sof_dfsentry *dfse = inode->i_private; |
| 119 | struct snd_sof_dev *sdev = dfse->sdev; |
| 120 | |
| 121 | /* avoid duplicate traces at next open */ |
| 122 | if (!sdev->dtrace_is_enabled) |
| 123 | sdev->host_offset = 0; |
| 124 | |
| 125 | return 0; |
| 126 | } |
| 127 | |
| 128 | static const struct file_operations sof_dfs_trace_fops = { |
| 129 | .open = simple_open, |
| 130 | .read = sof_dfsentry_trace_read, |
| 131 | .llseek = default_llseek, |
| 132 | .release = sof_dfsentry_trace_release, |
| 133 | }; |
| 134 | |
| 135 | static int trace_debugfs_create(struct snd_sof_dev *sdev) |
| 136 | { |
| 137 | struct snd_sof_dfsentry *dfse; |
| 138 | |
| 139 | if (!sdev) |
| 140 | return -EINVAL; |
| 141 | |
| 142 | dfse = devm_kzalloc(sdev->dev, sizeof(*dfse), GFP_KERNEL); |
| 143 | if (!dfse) |
| 144 | return -ENOMEM; |
| 145 | |
| 146 | dfse->type = SOF_DFSENTRY_TYPE_BUF; |
| 147 | dfse->buf = sdev->dmatb.area; |
| 148 | dfse->size = sdev->dmatb.bytes; |
| 149 | dfse->sdev = sdev; |
| 150 | |
| 151 | debugfs_create_file("trace", 0444, sdev->debugfs_root, dfse, |
| 152 | &sof_dfs_trace_fops); |
| 153 | |
| 154 | return 0; |
| 155 | } |
| 156 | |
| 157 | int snd_sof_init_trace_ipc(struct snd_sof_dev *sdev) |
| 158 | { |
| 159 | struct sof_ipc_fw_ready *ready = &sdev->fw_ready; |
| 160 | struct sof_ipc_fw_version *v = &ready->version; |
| 161 | struct sof_ipc_dma_trace_params_ext params; |
| 162 | struct sof_ipc_reply ipc_reply; |
| 163 | int ret; |
| 164 | |
| 165 | if (sdev->dtrace_is_enabled || !sdev->dma_trace_pages) |
| 166 | return -EINVAL; |
| 167 | |
| 168 | /* set IPC parameters */ |
| 169 | params.hdr.cmd = SOF_IPC_GLB_TRACE_MSG; |
| 170 | /* PARAMS_EXT is only supported from ABI 3.7.0 onwards */ |
| 171 | if (v->abi_version >= SOF_ABI_VER(3, 7, 0)) { |
| 172 | params.hdr.size = sizeof(struct sof_ipc_dma_trace_params_ext); |
| 173 | params.hdr.cmd |= SOF_IPC_TRACE_DMA_PARAMS_EXT; |
| 174 | params.timestamp_ns = ktime_get(); /* in nanosecond */ |
| 175 | } else { |
| 176 | params.hdr.size = sizeof(struct sof_ipc_dma_trace_params); |
| 177 | params.hdr.cmd |= SOF_IPC_TRACE_DMA_PARAMS; |
| 178 | } |
| 179 | params.buffer.phy_addr = sdev->dmatp.addr; |
| 180 | params.buffer.size = sdev->dmatb.bytes; |
| 181 | params.buffer.pages = sdev->dma_trace_pages; |
| 182 | params.stream_tag = 0; |
| 183 | |
| 184 | sdev->host_offset = 0; |
| 185 | sdev->dtrace_draining = false; |
| 186 | |
| 187 | ret = snd_sof_dma_trace_init(sdev, ¶ms.stream_tag); |
| 188 | if (ret < 0) { |
| 189 | dev_err(sdev->dev, |
| 190 | "error: fail in snd_sof_dma_trace_init %d\n", ret); |
| 191 | return ret; |
| 192 | } |
| 193 | dev_dbg(sdev->dev, "stream_tag: %d\n", params.stream_tag); |
| 194 | |
| 195 | /* send IPC to the DSP */ |
| 196 | ret = sof_ipc_tx_message(sdev->ipc, |
| 197 | params.hdr.cmd, ¶ms, sizeof(params), |
| 198 | &ipc_reply, sizeof(ipc_reply)); |
| 199 | if (ret < 0) { |
| 200 | dev_err(sdev->dev, |
| 201 | "error: can't set params for DMA for trace %d\n", ret); |
| 202 | goto trace_release; |
| 203 | } |
| 204 | |
| 205 | ret = snd_sof_dma_trace_trigger(sdev, SNDRV_PCM_TRIGGER_START); |
| 206 | if (ret < 0) { |
| 207 | dev_err(sdev->dev, |
| 208 | "error: snd_sof_dma_trace_trigger: start: %d\n", ret); |
| 209 | goto trace_release; |
| 210 | } |
| 211 | |
| 212 | sdev->dtrace_is_enabled = true; |
| 213 | |
| 214 | return 0; |
| 215 | |
| 216 | trace_release: |
| 217 | snd_sof_dma_trace_release(sdev); |
| 218 | return ret; |
| 219 | } |
| 220 | |
| 221 | int snd_sof_init_trace(struct snd_sof_dev *sdev) |
| 222 | { |
| 223 | int ret; |
| 224 | |
| 225 | /* set false before start initialization */ |
| 226 | sdev->dtrace_is_enabled = false; |
| 227 | |
| 228 | /* allocate trace page table buffer */ |
| 229 | ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV, sdev->dev, |
| 230 | PAGE_SIZE, &sdev->dmatp); |
| 231 | if (ret < 0) { |
| 232 | dev_err(sdev->dev, |
| 233 | "error: can't alloc page table for trace %d\n", ret); |
| 234 | return ret; |
| 235 | } |
| 236 | |
| 237 | /* allocate trace data buffer */ |
| 238 | ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV_SG, sdev->dev, |
| 239 | DMA_BUF_SIZE_FOR_TRACE, &sdev->dmatb); |
| 240 | if (ret < 0) { |
| 241 | dev_err(sdev->dev, |
| 242 | "error: can't alloc buffer for trace %d\n", ret); |
| 243 | goto page_err; |
| 244 | } |
| 245 | |
| 246 | /* create compressed page table for audio firmware */ |
| 247 | ret = snd_sof_create_page_table(sdev, &sdev->dmatb, sdev->dmatp.area, |
| 248 | sdev->dmatb.bytes); |
| 249 | if (ret < 0) |
| 250 | goto table_err; |
| 251 | |
| 252 | sdev->dma_trace_pages = ret; |
| 253 | dev_dbg(sdev->dev, "dma_trace_pages: %d\n", sdev->dma_trace_pages); |
| 254 | |
| 255 | if (sdev->first_boot) { |
| 256 | ret = trace_debugfs_create(sdev); |
| 257 | if (ret < 0) |
| 258 | goto table_err; |
| 259 | } |
| 260 | |
| 261 | init_waitqueue_head(&sdev->trace_sleep); |
| 262 | |
| 263 | ret = snd_sof_init_trace_ipc(sdev); |
| 264 | if (ret < 0) |
| 265 | goto table_err; |
| 266 | |
| 267 | return 0; |
| 268 | table_err: |
| 269 | sdev->dma_trace_pages = 0; |
| 270 | snd_dma_free_pages(&sdev->dmatb); |
| 271 | page_err: |
| 272 | snd_dma_free_pages(&sdev->dmatp); |
| 273 | return ret; |
| 274 | } |
| 275 | EXPORT_SYMBOL(snd_sof_init_trace); |
| 276 | |
| 277 | int snd_sof_trace_update_pos(struct snd_sof_dev *sdev, |
| 278 | struct sof_ipc_dma_trace_posn *posn) |
| 279 | { |
| 280 | if (sdev->dtrace_is_enabled && sdev->host_offset != posn->host_offset) { |
| 281 | sdev->host_offset = posn->host_offset; |
| 282 | wake_up(&sdev->trace_sleep); |
| 283 | } |
| 284 | |
| 285 | if (posn->overflow != 0) |
| 286 | dev_err(sdev->dev, |
| 287 | "error: DSP trace buffer overflow %u bytes. Total messages %d\n", |
| 288 | posn->overflow, posn->messages); |
| 289 | |
| 290 | return 0; |
| 291 | } |
| 292 | |
| 293 | /* an error has occurred within the DSP that prevents further trace */ |
| 294 | void snd_sof_trace_notify_for_error(struct snd_sof_dev *sdev) |
| 295 | { |
| 296 | if (sdev->dtrace_is_enabled) { |
| 297 | dev_err(sdev->dev, "error: waking up any trace sleepers\n"); |
| 298 | sdev->dtrace_error = true; |
| 299 | wake_up(&sdev->trace_sleep); |
| 300 | } |
| 301 | } |
| 302 | EXPORT_SYMBOL(snd_sof_trace_notify_for_error); |
| 303 | |
| 304 | void snd_sof_release_trace(struct snd_sof_dev *sdev) |
| 305 | { |
| 306 | int ret; |
| 307 | |
| 308 | if (!sdev->dtrace_is_enabled) |
| 309 | return; |
| 310 | |
| 311 | ret = snd_sof_dma_trace_trigger(sdev, SNDRV_PCM_TRIGGER_STOP); |
| 312 | if (ret < 0) |
| 313 | dev_err(sdev->dev, |
| 314 | "error: snd_sof_dma_trace_trigger: stop: %d\n", ret); |
| 315 | |
| 316 | ret = snd_sof_dma_trace_release(sdev); |
| 317 | if (ret < 0) |
| 318 | dev_err(sdev->dev, |
| 319 | "error: fail in snd_sof_dma_trace_release %d\n", ret); |
| 320 | |
| 321 | sdev->dtrace_is_enabled = false; |
| 322 | sdev->dtrace_draining = true; |
| 323 | wake_up(&sdev->trace_sleep); |
| 324 | } |
| 325 | EXPORT_SYMBOL(snd_sof_release_trace); |
| 326 | |
| 327 | void snd_sof_free_trace(struct snd_sof_dev *sdev) |
| 328 | { |
| 329 | snd_sof_release_trace(sdev); |
| 330 | |
| 331 | snd_dma_free_pages(&sdev->dmatb); |
| 332 | snd_dma_free_pages(&sdev->dmatp); |
| 333 | } |
| 334 | EXPORT_SYMBOL(snd_sof_free_trace); |