1 | // SPDX-License-Identifier: GPL-2.0-or-later |
2 | /* |
3 | * Digital Audio (PCM) abstract layer |
4 | * Copyright (c) by Jaroslav Kysela <perex@perex.cz> |
5 | */ |
6 | |
7 | #include <linux/compat.h> |
8 | #include <linux/mm.h> |
9 | #include <linux/module.h> |
10 | #include <linux/file.h> |
11 | #include <linux/slab.h> |
12 | #include <linux/sched/signal.h> |
13 | #include <linux/time.h> |
14 | #include <linux/pm_qos.h> |
15 | #include <linux/io.h> |
16 | #include <linux/dma-mapping.h> |
17 | #include <linux/vmalloc.h> |
18 | #include <sound/core.h> |
19 | #include <sound/control.h> |
20 | #include <sound/info.h> |
21 | #include <sound/pcm.h> |
22 | #include <sound/pcm_params.h> |
23 | #include <sound/timer.h> |
24 | #include <sound/minors.h> |
25 | #include <linux/uio.h> |
26 | #include <linux/delay.h> |
27 | |
28 | #include "pcm_local.h" |
29 | |
30 | #ifdef CONFIG_SND_DEBUG |
31 | #define CREATE_TRACE_POINTS |
32 | #include "pcm_param_trace.h" |
33 | #else |
34 | #define trace_hw_mask_param_enabled() 0 |
35 | #define trace_hw_interval_param_enabled() 0 |
36 | #define trace_hw_mask_param(substream, type, index, prev, curr) |
37 | #define trace_hw_interval_param(substream, type, index, prev, curr) |
38 | #endif |
39 | |
40 | /* |
41 | * Compatibility |
42 | */ |
43 | |
44 | struct snd_pcm_hw_params_old { |
45 | unsigned int flags; |
46 | unsigned int masks[SNDRV_PCM_HW_PARAM_SUBFORMAT - |
47 | SNDRV_PCM_HW_PARAM_ACCESS + 1]; |
48 | struct snd_interval intervals[SNDRV_PCM_HW_PARAM_TICK_TIME - |
49 | SNDRV_PCM_HW_PARAM_SAMPLE_BITS + 1]; |
50 | unsigned int rmask; |
51 | unsigned int cmask; |
52 | unsigned int info; |
53 | unsigned int msbits; |
54 | unsigned int rate_num; |
55 | unsigned int rate_den; |
56 | snd_pcm_uframes_t fifo_size; |
57 | unsigned char reserved[64]; |
58 | }; |
59 | |
60 | #ifdef CONFIG_SND_SUPPORT_OLD_API |
61 | #define SNDRV_PCM_IOCTL_HW_REFINE_OLD _IOWR('A', 0x10, struct snd_pcm_hw_params_old) |
62 | #define SNDRV_PCM_IOCTL_HW_PARAMS_OLD _IOWR('A', 0x11, struct snd_pcm_hw_params_old) |
63 | |
64 | static int snd_pcm_hw_refine_old_user(struct snd_pcm_substream *substream, |
65 | struct snd_pcm_hw_params_old __user * _oparams); |
66 | static int snd_pcm_hw_params_old_user(struct snd_pcm_substream *substream, |
67 | struct snd_pcm_hw_params_old __user * _oparams); |
68 | #endif |
69 | static int snd_pcm_open(struct file *file, struct snd_pcm *pcm, int stream); |
70 | |
71 | /* |
72 | * |
73 | */ |
74 | |
75 | static DECLARE_RWSEM(snd_pcm_link_rwsem); |
76 | |
77 | void snd_pcm_group_init(struct snd_pcm_group *group) |
78 | { |
79 | spin_lock_init(&group->lock); |
80 | mutex_init(&group->mutex); |
81 | INIT_LIST_HEAD(list: &group->substreams); |
82 | refcount_set(r: &group->refs, n: 1); |
83 | } |
84 | |
85 | /* define group lock helpers */ |
86 | #define DEFINE_PCM_GROUP_LOCK(action, mutex_action) \ |
87 | static void snd_pcm_group_ ## action(struct snd_pcm_group *group, bool nonatomic) \ |
88 | { \ |
89 | if (nonatomic) \ |
90 | mutex_ ## mutex_action(&group->mutex); \ |
91 | else \ |
92 | spin_ ## action(&group->lock); \ |
93 | } |
94 | |
95 | DEFINE_PCM_GROUP_LOCK(lock, lock); |
96 | DEFINE_PCM_GROUP_LOCK(unlock, unlock); |
97 | DEFINE_PCM_GROUP_LOCK(lock_irq, lock); |
98 | DEFINE_PCM_GROUP_LOCK(unlock_irq, unlock); |
99 | |
100 | /** |
101 | * snd_pcm_stream_lock - Lock the PCM stream |
102 | * @substream: PCM substream |
103 | * |
104 | * This locks the PCM stream's spinlock or mutex depending on the nonatomic |
105 | * flag of the given substream. This also takes the global link rw lock |
106 | * (or rw sem), too, for avoiding the race with linked streams. |
107 | */ |
108 | void snd_pcm_stream_lock(struct snd_pcm_substream *substream) |
109 | { |
110 | snd_pcm_group_lock(group: &substream->self_group, nonatomic: substream->pcm->nonatomic); |
111 | } |
112 | EXPORT_SYMBOL_GPL(snd_pcm_stream_lock); |
113 | |
114 | /** |
115 | * snd_pcm_stream_unlock - Unlock the PCM stream |
116 | * @substream: PCM substream |
117 | * |
118 | * This unlocks the PCM stream that has been locked via snd_pcm_stream_lock(). |
119 | */ |
120 | void snd_pcm_stream_unlock(struct snd_pcm_substream *substream) |
121 | { |
122 | snd_pcm_group_unlock(group: &substream->self_group, nonatomic: substream->pcm->nonatomic); |
123 | } |
124 | EXPORT_SYMBOL_GPL(snd_pcm_stream_unlock); |
125 | |
126 | /** |
127 | * snd_pcm_stream_lock_irq - Lock the PCM stream |
128 | * @substream: PCM substream |
129 | * |
130 | * This locks the PCM stream like snd_pcm_stream_lock() and disables the local |
131 | * IRQ (only when nonatomic is false). In nonatomic case, this is identical |
132 | * as snd_pcm_stream_lock(). |
133 | */ |
134 | void snd_pcm_stream_lock_irq(struct snd_pcm_substream *substream) |
135 | { |
136 | snd_pcm_group_lock_irq(group: &substream->self_group, |
137 | nonatomic: substream->pcm->nonatomic); |
138 | } |
139 | EXPORT_SYMBOL_GPL(snd_pcm_stream_lock_irq); |
140 | |
141 | static void snd_pcm_stream_lock_nested(struct snd_pcm_substream *substream) |
142 | { |
143 | struct snd_pcm_group *group = &substream->self_group; |
144 | |
145 | if (substream->pcm->nonatomic) |
146 | mutex_lock_nested(lock: &group->mutex, SINGLE_DEPTH_NESTING); |
147 | else |
148 | spin_lock_nested(&group->lock, SINGLE_DEPTH_NESTING); |
149 | } |
150 | |
151 | /** |
152 | * snd_pcm_stream_unlock_irq - Unlock the PCM stream |
153 | * @substream: PCM substream |
154 | * |
155 | * This is a counter-part of snd_pcm_stream_lock_irq(). |
156 | */ |
157 | void snd_pcm_stream_unlock_irq(struct snd_pcm_substream *substream) |
158 | { |
159 | snd_pcm_group_unlock_irq(group: &substream->self_group, |
160 | nonatomic: substream->pcm->nonatomic); |
161 | } |
162 | EXPORT_SYMBOL_GPL(snd_pcm_stream_unlock_irq); |
163 | |
164 | unsigned long _snd_pcm_stream_lock_irqsave(struct snd_pcm_substream *substream) |
165 | { |
166 | unsigned long flags = 0; |
167 | if (substream->pcm->nonatomic) |
168 | mutex_lock(&substream->self_group.mutex); |
169 | else |
170 | spin_lock_irqsave(&substream->self_group.lock, flags); |
171 | return flags; |
172 | } |
173 | EXPORT_SYMBOL_GPL(_snd_pcm_stream_lock_irqsave); |
174 | |
175 | unsigned long _snd_pcm_stream_lock_irqsave_nested(struct snd_pcm_substream *substream) |
176 | { |
177 | unsigned long flags = 0; |
178 | if (substream->pcm->nonatomic) |
179 | mutex_lock_nested(lock: &substream->self_group.mutex, |
180 | SINGLE_DEPTH_NESTING); |
181 | else |
182 | spin_lock_irqsave_nested(&substream->self_group.lock, flags, |
183 | SINGLE_DEPTH_NESTING); |
184 | return flags; |
185 | } |
186 | EXPORT_SYMBOL_GPL(_snd_pcm_stream_lock_irqsave_nested); |
187 | |
188 | /** |
189 | * snd_pcm_stream_unlock_irqrestore - Unlock the PCM stream |
190 | * @substream: PCM substream |
191 | * @flags: irq flags |
192 | * |
193 | * This is a counter-part of snd_pcm_stream_lock_irqsave(). |
194 | */ |
195 | void snd_pcm_stream_unlock_irqrestore(struct snd_pcm_substream *substream, |
196 | unsigned long flags) |
197 | { |
198 | if (substream->pcm->nonatomic) |
199 | mutex_unlock(lock: &substream->self_group.mutex); |
200 | else |
201 | spin_unlock_irqrestore(lock: &substream->self_group.lock, flags); |
202 | } |
203 | EXPORT_SYMBOL_GPL(snd_pcm_stream_unlock_irqrestore); |
204 | |
205 | /* Run PCM ioctl ops */ |
206 | static int snd_pcm_ops_ioctl(struct snd_pcm_substream *substream, |
207 | unsigned cmd, void *arg) |
208 | { |
209 | if (substream->ops->ioctl) |
210 | return substream->ops->ioctl(substream, cmd, arg); |
211 | else |
212 | return snd_pcm_lib_ioctl(substream, cmd, arg); |
213 | } |
214 | |
215 | int snd_pcm_info(struct snd_pcm_substream *substream, struct snd_pcm_info *info) |
216 | { |
217 | struct snd_pcm *pcm = substream->pcm; |
218 | struct snd_pcm_str *pstr = substream->pstr; |
219 | |
220 | memset(info, 0, sizeof(*info)); |
221 | info->card = pcm->card->number; |
222 | info->device = pcm->device; |
223 | info->stream = substream->stream; |
224 | info->subdevice = substream->number; |
225 | strscpy(info->id, pcm->id, sizeof(info->id)); |
226 | strscpy(info->name, pcm->name, sizeof(info->name)); |
227 | info->dev_class = pcm->dev_class; |
228 | info->dev_subclass = pcm->dev_subclass; |
229 | info->subdevices_count = pstr->substream_count; |
230 | info->subdevices_avail = pstr->substream_count - pstr->substream_opened; |
231 | strscpy(info->subname, substream->name, sizeof(info->subname)); |
232 | |
233 | return 0; |
234 | } |
235 | |
236 | int snd_pcm_info_user(struct snd_pcm_substream *substream, |
237 | struct snd_pcm_info __user * _info) |
238 | { |
239 | struct snd_pcm_info *info __free(kfree) = NULL; |
240 | int err; |
241 | |
242 | info = kmalloc(size: sizeof(*info), GFP_KERNEL); |
243 | if (! info) |
244 | return -ENOMEM; |
245 | err = snd_pcm_info(substream, info); |
246 | if (err >= 0) { |
247 | if (copy_to_user(to: _info, from: info, n: sizeof(*info))) |
248 | err = -EFAULT; |
249 | } |
250 | return err; |
251 | } |
252 | |
253 | /* macro for simplified cast */ |
254 | #define PARAM_MASK_BIT(b) (1U << (__force int)(b)) |
255 | |
256 | static bool hw_support_mmap(struct snd_pcm_substream *substream) |
257 | { |
258 | struct snd_dma_buffer *dmabuf; |
259 | |
260 | if (!(substream->runtime->hw.info & SNDRV_PCM_INFO_MMAP)) |
261 | return false; |
262 | |
263 | if (substream->ops->mmap || substream->ops->page) |
264 | return true; |
265 | |
266 | dmabuf = snd_pcm_get_dma_buf(substream); |
267 | if (!dmabuf) |
268 | dmabuf = &substream->dma_buffer; |
269 | switch (dmabuf->dev.type) { |
270 | case SNDRV_DMA_TYPE_UNKNOWN: |
271 | /* we can't know the device, so just assume that the driver does |
272 | * everything right |
273 | */ |
274 | return true; |
275 | case SNDRV_DMA_TYPE_CONTINUOUS: |
276 | case SNDRV_DMA_TYPE_VMALLOC: |
277 | return true; |
278 | default: |
279 | return dma_can_mmap(dev: dmabuf->dev.dev); |
280 | } |
281 | } |
282 | |
283 | static int constrain_mask_params(struct snd_pcm_substream *substream, |
284 | struct snd_pcm_hw_params *params) |
285 | { |
286 | struct snd_pcm_hw_constraints *constrs = |
287 | &substream->runtime->hw_constraints; |
288 | struct snd_mask *m; |
289 | unsigned int k; |
290 | struct snd_mask old_mask __maybe_unused; |
291 | int changed; |
292 | |
293 | for (k = SNDRV_PCM_HW_PARAM_FIRST_MASK; k <= SNDRV_PCM_HW_PARAM_LAST_MASK; k++) { |
294 | m = hw_param_mask(params, var: k); |
295 | if (snd_mask_empty(mask: m)) |
296 | return -EINVAL; |
297 | |
298 | /* This parameter is not requested to change by a caller. */ |
299 | if (!(params->rmask & PARAM_MASK_BIT(k))) |
300 | continue; |
301 | |
302 | if (trace_hw_mask_param_enabled()) |
303 | old_mask = *m; |
304 | |
305 | changed = snd_mask_refine(mask: m, v: constrs_mask(constrs, var: k)); |
306 | if (changed < 0) |
307 | return changed; |
308 | if (changed == 0) |
309 | continue; |
310 | |
311 | /* Set corresponding flag so that the caller gets it. */ |
312 | trace_hw_mask_param(substream, type: k, index: 0, prev: &old_mask, curr: m); |
313 | params->cmask |= PARAM_MASK_BIT(k); |
314 | } |
315 | |
316 | return 0; |
317 | } |
318 | |
319 | static int constrain_interval_params(struct snd_pcm_substream *substream, |
320 | struct snd_pcm_hw_params *params) |
321 | { |
322 | struct snd_pcm_hw_constraints *constrs = |
323 | &substream->runtime->hw_constraints; |
324 | struct snd_interval *i; |
325 | unsigned int k; |
326 | struct snd_interval old_interval __maybe_unused; |
327 | int changed; |
328 | |
329 | for (k = SNDRV_PCM_HW_PARAM_FIRST_INTERVAL; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++) { |
330 | i = hw_param_interval(params, var: k); |
331 | if (snd_interval_empty(i)) |
332 | return -EINVAL; |
333 | |
334 | /* This parameter is not requested to change by a caller. */ |
335 | if (!(params->rmask & PARAM_MASK_BIT(k))) |
336 | continue; |
337 | |
338 | if (trace_hw_interval_param_enabled()) |
339 | old_interval = *i; |
340 | |
341 | changed = snd_interval_refine(i, v: constrs_interval(constrs, var: k)); |
342 | if (changed < 0) |
343 | return changed; |
344 | if (changed == 0) |
345 | continue; |
346 | |
347 | /* Set corresponding flag so that the caller gets it. */ |
348 | trace_hw_interval_param(substream, type: k, index: 0, prev: &old_interval, curr: i); |
349 | params->cmask |= PARAM_MASK_BIT(k); |
350 | } |
351 | |
352 | return 0; |
353 | } |
354 | |
355 | static int constrain_params_by_rules(struct snd_pcm_substream *substream, |
356 | struct snd_pcm_hw_params *params) |
357 | { |
358 | struct snd_pcm_hw_constraints *constrs = |
359 | &substream->runtime->hw_constraints; |
360 | unsigned int k; |
361 | unsigned int *rstamps __free(kfree) = NULL; |
362 | unsigned int vstamps[SNDRV_PCM_HW_PARAM_LAST_INTERVAL + 1]; |
363 | unsigned int stamp; |
364 | struct snd_pcm_hw_rule *r; |
365 | unsigned int d; |
366 | struct snd_mask old_mask __maybe_unused; |
367 | struct snd_interval old_interval __maybe_unused; |
368 | bool again; |
369 | int changed, err = 0; |
370 | |
371 | /* |
372 | * Each application of rule has own sequence number. |
373 | * |
374 | * Each member of 'rstamps' array represents the sequence number of |
375 | * recent application of corresponding rule. |
376 | */ |
377 | rstamps = kcalloc(n: constrs->rules_num, size: sizeof(unsigned int), GFP_KERNEL); |
378 | if (!rstamps) |
379 | return -ENOMEM; |
380 | |
381 | /* |
382 | * Each member of 'vstamps' array represents the sequence number of |
383 | * recent application of rule in which corresponding parameters were |
384 | * changed. |
385 | * |
386 | * In initial state, elements corresponding to parameters requested by |
387 | * a caller is 1. For unrequested parameters, corresponding members |
388 | * have 0 so that the parameters are never changed anymore. |
389 | */ |
390 | for (k = 0; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++) |
391 | vstamps[k] = (params->rmask & PARAM_MASK_BIT(k)) ? 1 : 0; |
392 | |
393 | /* Due to the above design, actual sequence number starts at 2. */ |
394 | stamp = 2; |
395 | retry: |
396 | /* Apply all rules in order. */ |
397 | again = false; |
398 | for (k = 0; k < constrs->rules_num; k++) { |
399 | r = &constrs->rules[k]; |
400 | |
401 | /* |
402 | * Check condition bits of this rule. When the rule has |
403 | * some condition bits, parameter without the bits is |
404 | * never processed. SNDRV_PCM_HW_PARAMS_NO_PERIOD_WAKEUP |
405 | * is an example of the condition bits. |
406 | */ |
407 | if (r->cond && !(r->cond & params->flags)) |
408 | continue; |
409 | |
410 | /* |
411 | * The 'deps' array includes maximum four dependencies |
412 | * to SNDRV_PCM_HW_PARAM_XXXs for this rule. The fifth |
413 | * member of this array is a sentinel and should be |
414 | * negative value. |
415 | * |
416 | * This rule should be processed in this time when dependent |
417 | * parameters were changed at former applications of the other |
418 | * rules. |
419 | */ |
420 | for (d = 0; r->deps[d] >= 0; d++) { |
421 | if (vstamps[r->deps[d]] > rstamps[k]) |
422 | break; |
423 | } |
424 | if (r->deps[d] < 0) |
425 | continue; |
426 | |
427 | if (trace_hw_mask_param_enabled()) { |
428 | if (hw_is_mask(var: r->var)) |
429 | old_mask = *hw_param_mask(params, var: r->var); |
430 | } |
431 | if (trace_hw_interval_param_enabled()) { |
432 | if (hw_is_interval(var: r->var)) |
433 | old_interval = *hw_param_interval(params, var: r->var); |
434 | } |
435 | |
436 | changed = r->func(params, r); |
437 | if (changed < 0) |
438 | return changed; |
439 | |
440 | /* |
441 | * When the parameter is changed, notify it to the caller |
442 | * by corresponding returned bit, then preparing for next |
443 | * iteration. |
444 | */ |
445 | if (changed && r->var >= 0) { |
446 | if (hw_is_mask(var: r->var)) { |
447 | trace_hw_mask_param(substream, type: r->var, |
448 | index: k + 1, prev: &old_mask, |
449 | curr: hw_param_mask(params, var: r->var)); |
450 | } |
451 | if (hw_is_interval(var: r->var)) { |
452 | trace_hw_interval_param(substream, type: r->var, |
453 | index: k + 1, prev: &old_interval, |
454 | curr: hw_param_interval(params, var: r->var)); |
455 | } |
456 | |
457 | params->cmask |= PARAM_MASK_BIT(r->var); |
458 | vstamps[r->var] = stamp; |
459 | again = true; |
460 | } |
461 | |
462 | rstamps[k] = stamp++; |
463 | } |
464 | |
465 | /* Iterate to evaluate all rules till no parameters are changed. */ |
466 | if (again) |
467 | goto retry; |
468 | |
469 | return err; |
470 | } |
471 | |
472 | static int fixup_unreferenced_params(struct snd_pcm_substream *substream, |
473 | struct snd_pcm_hw_params *params) |
474 | { |
475 | const struct snd_interval *i; |
476 | const struct snd_mask *m; |
477 | struct snd_mask *m_rw; |
478 | int err; |
479 | |
480 | if (!params->msbits) { |
481 | i = hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_SAMPLE_BITS); |
482 | if (snd_interval_single(i)) |
483 | params->msbits = snd_interval_value(i); |
484 | m = hw_param_mask_c(params, SNDRV_PCM_HW_PARAM_FORMAT); |
485 | if (snd_mask_single(mask: m)) { |
486 | snd_pcm_format_t format = (__force snd_pcm_format_t)snd_mask_min(mask: m); |
487 | params->msbits = snd_pcm_format_width(format); |
488 | } |
489 | } |
490 | |
491 | if (params->msbits) { |
492 | m = hw_param_mask_c(params, SNDRV_PCM_HW_PARAM_FORMAT); |
493 | if (snd_mask_single(mask: m)) { |
494 | snd_pcm_format_t format = (__force snd_pcm_format_t)snd_mask_min(mask: m); |
495 | |
496 | if (snd_pcm_format_linear(format) && |
497 | snd_pcm_format_width(format) != params->msbits) { |
498 | m_rw = hw_param_mask(params, SNDRV_PCM_HW_PARAM_SUBFORMAT); |
499 | snd_mask_reset(mask: m_rw, |
500 | val: (__force unsigned)SNDRV_PCM_SUBFORMAT_MSBITS_MAX); |
501 | if (snd_mask_empty(mask: m_rw)) |
502 | return -EINVAL; |
503 | } |
504 | } |
505 | } |
506 | |
507 | if (!params->rate_den) { |
508 | i = hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_RATE); |
509 | if (snd_interval_single(i)) { |
510 | params->rate_num = snd_interval_value(i); |
511 | params->rate_den = 1; |
512 | } |
513 | } |
514 | |
515 | if (!params->fifo_size) { |
516 | m = hw_param_mask_c(params, SNDRV_PCM_HW_PARAM_FORMAT); |
517 | i = hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_CHANNELS); |
518 | if (snd_mask_single(mask: m) && snd_interval_single(i)) { |
519 | err = snd_pcm_ops_ioctl(substream, |
520 | SNDRV_PCM_IOCTL1_FIFO_SIZE, |
521 | arg: params); |
522 | if (err < 0) |
523 | return err; |
524 | } |
525 | } |
526 | |
527 | if (!params->info) { |
528 | params->info = substream->runtime->hw.info; |
529 | params->info &= ~(SNDRV_PCM_INFO_FIFO_IN_FRAMES | |
530 | SNDRV_PCM_INFO_DRAIN_TRIGGER); |
531 | if (!hw_support_mmap(substream)) |
532 | params->info &= ~(SNDRV_PCM_INFO_MMAP | |
533 | SNDRV_PCM_INFO_MMAP_VALID); |
534 | } |
535 | |
536 | return 0; |
537 | } |
538 | |
539 | int snd_pcm_hw_refine(struct snd_pcm_substream *substream, |
540 | struct snd_pcm_hw_params *params) |
541 | { |
542 | int err; |
543 | |
544 | params->info = 0; |
545 | params->fifo_size = 0; |
546 | if (params->rmask & PARAM_MASK_BIT(SNDRV_PCM_HW_PARAM_SAMPLE_BITS)) |
547 | params->msbits = 0; |
548 | if (params->rmask & PARAM_MASK_BIT(SNDRV_PCM_HW_PARAM_RATE)) { |
549 | params->rate_num = 0; |
550 | params->rate_den = 0; |
551 | } |
552 | |
553 | err = constrain_mask_params(substream, params); |
554 | if (err < 0) |
555 | return err; |
556 | |
557 | err = constrain_interval_params(substream, params); |
558 | if (err < 0) |
559 | return err; |
560 | |
561 | err = constrain_params_by_rules(substream, params); |
562 | if (err < 0) |
563 | return err; |
564 | |
565 | params->rmask = 0; |
566 | |
567 | return 0; |
568 | } |
569 | EXPORT_SYMBOL(snd_pcm_hw_refine); |
570 | |
571 | static int snd_pcm_hw_refine_user(struct snd_pcm_substream *substream, |
572 | struct snd_pcm_hw_params __user * _params) |
573 | { |
574 | struct snd_pcm_hw_params *params __free(kfree) = NULL; |
575 | int err; |
576 | |
577 | params = memdup_user(_params, sizeof(*params)); |
578 | if (IS_ERR(ptr: params)) |
579 | return PTR_ERR(no_free_ptr(params)); |
580 | |
581 | err = snd_pcm_hw_refine(substream, params); |
582 | if (err < 0) |
583 | return err; |
584 | |
585 | err = fixup_unreferenced_params(substream, params); |
586 | if (err < 0) |
587 | return err; |
588 | |
589 | if (copy_to_user(to: _params, from: params, n: sizeof(*params))) |
590 | return -EFAULT; |
591 | return 0; |
592 | } |
593 | |
594 | static int period_to_usecs(struct snd_pcm_runtime *runtime) |
595 | { |
596 | int usecs; |
597 | |
598 | if (! runtime->rate) |
599 | return -1; /* invalid */ |
600 | |
601 | /* take 75% of period time as the deadline */ |
602 | usecs = (750000 / runtime->rate) * runtime->period_size; |
603 | usecs += ((750000 % runtime->rate) * runtime->period_size) / |
604 | runtime->rate; |
605 | |
606 | return usecs; |
607 | } |
608 | |
609 | static void snd_pcm_set_state(struct snd_pcm_substream *substream, |
610 | snd_pcm_state_t state) |
611 | { |
612 | guard(pcm_stream_lock_irq)(l: substream); |
613 | if (substream->runtime->state != SNDRV_PCM_STATE_DISCONNECTED) |
614 | __snd_pcm_set_state(runtime: substream->runtime, state); |
615 | } |
616 | |
617 | static inline void snd_pcm_timer_notify(struct snd_pcm_substream *substream, |
618 | int event) |
619 | { |
620 | #ifdef CONFIG_SND_PCM_TIMER |
621 | if (substream->timer) |
622 | snd_timer_notify(timer: substream->timer, event, |
623 | tstamp: &substream->runtime->trigger_tstamp); |
624 | #endif |
625 | } |
626 | |
627 | void snd_pcm_sync_stop(struct snd_pcm_substream *substream, bool sync_irq) |
628 | { |
629 | if (substream->runtime && substream->runtime->stop_operating) { |
630 | substream->runtime->stop_operating = false; |
631 | if (substream->ops && substream->ops->sync_stop) |
632 | substream->ops->sync_stop(substream); |
633 | else if (sync_irq && substream->pcm->card->sync_irq > 0) |
634 | synchronize_irq(irq: substream->pcm->card->sync_irq); |
635 | } |
636 | } |
637 | |
638 | /** |
639 | * snd_pcm_hw_params_choose - choose a configuration defined by @params |
640 | * @pcm: PCM instance |
641 | * @params: the hw_params instance |
642 | * |
643 | * Choose one configuration from configuration space defined by @params. |
644 | * The configuration chosen is that obtained fixing in this order: |
645 | * first access, first format, first subformat, min channels, |
646 | * min rate, min period time, max buffer size, min tick time |
647 | * |
648 | * Return: Zero if successful, or a negative error code on failure. |
649 | */ |
650 | static int snd_pcm_hw_params_choose(struct snd_pcm_substream *pcm, |
651 | struct snd_pcm_hw_params *params) |
652 | { |
653 | static const int vars[] = { |
654 | SNDRV_PCM_HW_PARAM_ACCESS, |
655 | SNDRV_PCM_HW_PARAM_FORMAT, |
656 | SNDRV_PCM_HW_PARAM_SUBFORMAT, |
657 | SNDRV_PCM_HW_PARAM_CHANNELS, |
658 | SNDRV_PCM_HW_PARAM_RATE, |
659 | SNDRV_PCM_HW_PARAM_PERIOD_TIME, |
660 | SNDRV_PCM_HW_PARAM_BUFFER_SIZE, |
661 | SNDRV_PCM_HW_PARAM_TICK_TIME, |
662 | -1 |
663 | }; |
664 | const int *v; |
665 | struct snd_mask old_mask __maybe_unused; |
666 | struct snd_interval old_interval __maybe_unused; |
667 | int changed; |
668 | |
669 | for (v = vars; *v != -1; v++) { |
670 | /* Keep old parameter to trace. */ |
671 | if (trace_hw_mask_param_enabled()) { |
672 | if (hw_is_mask(var: *v)) |
673 | old_mask = *hw_param_mask(params, var: *v); |
674 | } |
675 | if (trace_hw_interval_param_enabled()) { |
676 | if (hw_is_interval(var: *v)) |
677 | old_interval = *hw_param_interval(params, var: *v); |
678 | } |
679 | if (*v != SNDRV_PCM_HW_PARAM_BUFFER_SIZE) |
680 | changed = snd_pcm_hw_param_first(pcm, params, var: *v, NULL); |
681 | else |
682 | changed = snd_pcm_hw_param_last(pcm, params, var: *v, NULL); |
683 | if (changed < 0) |
684 | return changed; |
685 | if (changed == 0) |
686 | continue; |
687 | |
688 | /* Trace the changed parameter. */ |
689 | if (hw_is_mask(var: *v)) { |
690 | trace_hw_mask_param(substream: pcm, type: *v, index: 0, prev: &old_mask, |
691 | curr: hw_param_mask(params, var: *v)); |
692 | } |
693 | if (hw_is_interval(var: *v)) { |
694 | trace_hw_interval_param(substream: pcm, type: *v, index: 0, prev: &old_interval, |
695 | curr: hw_param_interval(params, var: *v)); |
696 | } |
697 | } |
698 | |
699 | return 0; |
700 | } |
701 | |
702 | /* acquire buffer_mutex; if it's in r/w operation, return -EBUSY, otherwise |
703 | * block the further r/w operations |
704 | */ |
705 | static int snd_pcm_buffer_access_lock(struct snd_pcm_runtime *runtime) |
706 | { |
707 | if (!atomic_dec_unless_positive(v: &runtime->buffer_accessing)) |
708 | return -EBUSY; |
709 | mutex_lock(&runtime->buffer_mutex); |
710 | return 0; /* keep buffer_mutex, unlocked by below */ |
711 | } |
712 | |
713 | /* release buffer_mutex and clear r/w access flag */ |
714 | static void snd_pcm_buffer_access_unlock(struct snd_pcm_runtime *runtime) |
715 | { |
716 | mutex_unlock(lock: &runtime->buffer_mutex); |
717 | atomic_inc(v: &runtime->buffer_accessing); |
718 | } |
719 | |
720 | #if IS_ENABLED(CONFIG_SND_PCM_OSS) |
721 | #define is_oss_stream(substream) ((substream)->oss.oss) |
722 | #else |
723 | #define is_oss_stream(substream) false |
724 | #endif |
725 | |
726 | static int snd_pcm_hw_params(struct snd_pcm_substream *substream, |
727 | struct snd_pcm_hw_params *params) |
728 | { |
729 | struct snd_pcm_runtime *runtime; |
730 | int err, usecs; |
731 | unsigned int bits; |
732 | snd_pcm_uframes_t frames; |
733 | |
734 | if (PCM_RUNTIME_CHECK(substream)) |
735 | return -ENXIO; |
736 | runtime = substream->runtime; |
737 | err = snd_pcm_buffer_access_lock(runtime); |
738 | if (err < 0) |
739 | return err; |
740 | scoped_guard(pcm_stream_lock_irq, substream) { |
741 | switch (runtime->state) { |
742 | case SNDRV_PCM_STATE_OPEN: |
743 | case SNDRV_PCM_STATE_SETUP: |
744 | case SNDRV_PCM_STATE_PREPARED: |
745 | if (!is_oss_stream(substream) && |
746 | atomic_read(v: &substream->mmap_count)) |
747 | err = -EBADFD; |
748 | break; |
749 | default: |
750 | err = -EBADFD; |
751 | break; |
752 | } |
753 | } |
754 | if (err) |
755 | goto unlock; |
756 | |
757 | snd_pcm_sync_stop(substream, sync_irq: true); |
758 | |
759 | params->rmask = ~0U; |
760 | err = snd_pcm_hw_refine(substream, params); |
761 | if (err < 0) |
762 | goto _error; |
763 | |
764 | err = snd_pcm_hw_params_choose(pcm: substream, params); |
765 | if (err < 0) |
766 | goto _error; |
767 | |
768 | err = fixup_unreferenced_params(substream, params); |
769 | if (err < 0) |
770 | goto _error; |
771 | |
772 | if (substream->managed_buffer_alloc) { |
773 | err = snd_pcm_lib_malloc_pages(substream, |
774 | size: params_buffer_bytes(p: params)); |
775 | if (err < 0) |
776 | goto _error; |
777 | runtime->buffer_changed = err > 0; |
778 | } |
779 | |
780 | if (substream->ops->hw_params != NULL) { |
781 | err = substream->ops->hw_params(substream, params); |
782 | if (err < 0) |
783 | goto _error; |
784 | } |
785 | |
786 | runtime->access = params_access(p: params); |
787 | runtime->format = params_format(p: params); |
788 | runtime->subformat = params_subformat(p: params); |
789 | runtime->channels = params_channels(p: params); |
790 | runtime->rate = params_rate(p: params); |
791 | runtime->period_size = params_period_size(p: params); |
792 | runtime->periods = params_periods(p: params); |
793 | runtime->buffer_size = params_buffer_size(p: params); |
794 | runtime->info = params->info; |
795 | runtime->rate_num = params->rate_num; |
796 | runtime->rate_den = params->rate_den; |
797 | runtime->no_period_wakeup = |
798 | (params->info & SNDRV_PCM_INFO_NO_PERIOD_WAKEUP) && |
799 | (params->flags & SNDRV_PCM_HW_PARAMS_NO_PERIOD_WAKEUP); |
800 | |
801 | bits = snd_pcm_format_physical_width(format: runtime->format); |
802 | runtime->sample_bits = bits; |
803 | bits *= runtime->channels; |
804 | runtime->frame_bits = bits; |
805 | frames = 1; |
806 | while (bits % 8 != 0) { |
807 | bits *= 2; |
808 | frames *= 2; |
809 | } |
810 | runtime->byte_align = bits / 8; |
811 | runtime->min_align = frames; |
812 | |
813 | /* Default sw params */ |
814 | runtime->tstamp_mode = SNDRV_PCM_TSTAMP_NONE; |
815 | runtime->period_step = 1; |
816 | runtime->control->avail_min = runtime->period_size; |
817 | runtime->start_threshold = 1; |
818 | runtime->stop_threshold = runtime->buffer_size; |
819 | runtime->silence_threshold = 0; |
820 | runtime->silence_size = 0; |
821 | runtime->boundary = runtime->buffer_size; |
822 | while (runtime->boundary * 2 <= LONG_MAX - runtime->buffer_size) |
823 | runtime->boundary *= 2; |
824 | |
825 | /* clear the buffer for avoiding possible kernel info leaks */ |
826 | if (runtime->dma_area && !substream->ops->copy) { |
827 | size_t size = runtime->dma_bytes; |
828 | |
829 | if (runtime->info & SNDRV_PCM_INFO_MMAP) |
830 | size = PAGE_ALIGN(size); |
831 | memset(runtime->dma_area, 0, size); |
832 | } |
833 | |
834 | snd_pcm_timer_resolution_change(substream); |
835 | snd_pcm_set_state(substream, SNDRV_PCM_STATE_SETUP); |
836 | |
837 | if (cpu_latency_qos_request_active(req: &substream->latency_pm_qos_req)) |
838 | cpu_latency_qos_remove_request(req: &substream->latency_pm_qos_req); |
839 | usecs = period_to_usecs(runtime); |
840 | if (usecs >= 0) |
841 | cpu_latency_qos_add_request(req: &substream->latency_pm_qos_req, |
842 | value: usecs); |
843 | err = 0; |
844 | _error: |
845 | if (err) { |
846 | /* hardware might be unusable from this time, |
847 | * so we force application to retry to set |
848 | * the correct hardware parameter settings |
849 | */ |
850 | snd_pcm_set_state(substream, SNDRV_PCM_STATE_OPEN); |
851 | if (substream->ops->hw_free != NULL) |
852 | substream->ops->hw_free(substream); |
853 | if (substream->managed_buffer_alloc) |
854 | snd_pcm_lib_free_pages(substream); |
855 | } |
856 | unlock: |
857 | snd_pcm_buffer_access_unlock(runtime); |
858 | return err; |
859 | } |
860 | |
861 | static int snd_pcm_hw_params_user(struct snd_pcm_substream *substream, |
862 | struct snd_pcm_hw_params __user * _params) |
863 | { |
864 | struct snd_pcm_hw_params *params __free(kfree) = NULL; |
865 | int err; |
866 | |
867 | params = memdup_user(_params, sizeof(*params)); |
868 | if (IS_ERR(ptr: params)) |
869 | return PTR_ERR(no_free_ptr(params)); |
870 | |
871 | err = snd_pcm_hw_params(substream, params); |
872 | if (err < 0) |
873 | return err; |
874 | |
875 | if (copy_to_user(to: _params, from: params, n: sizeof(*params))) |
876 | return -EFAULT; |
877 | return err; |
878 | } |
879 | |
880 | static int do_hw_free(struct snd_pcm_substream *substream) |
881 | { |
882 | int result = 0; |
883 | |
884 | snd_pcm_sync_stop(substream, sync_irq: true); |
885 | if (substream->ops->hw_free) |
886 | result = substream->ops->hw_free(substream); |
887 | if (substream->managed_buffer_alloc) |
888 | snd_pcm_lib_free_pages(substream); |
889 | return result; |
890 | } |
891 | |
892 | static int snd_pcm_hw_free(struct snd_pcm_substream *substream) |
893 | { |
894 | struct snd_pcm_runtime *runtime; |
895 | int result = 0; |
896 | |
897 | if (PCM_RUNTIME_CHECK(substream)) |
898 | return -ENXIO; |
899 | runtime = substream->runtime; |
900 | result = snd_pcm_buffer_access_lock(runtime); |
901 | if (result < 0) |
902 | return result; |
903 | scoped_guard(pcm_stream_lock_irq, substream) { |
904 | switch (runtime->state) { |
905 | case SNDRV_PCM_STATE_SETUP: |
906 | case SNDRV_PCM_STATE_PREPARED: |
907 | if (atomic_read(v: &substream->mmap_count)) |
908 | result = -EBADFD; |
909 | break; |
910 | default: |
911 | result = -EBADFD; |
912 | break; |
913 | } |
914 | } |
915 | if (result) |
916 | goto unlock; |
917 | result = do_hw_free(substream); |
918 | snd_pcm_set_state(substream, SNDRV_PCM_STATE_OPEN); |
919 | cpu_latency_qos_remove_request(req: &substream->latency_pm_qos_req); |
920 | unlock: |
921 | snd_pcm_buffer_access_unlock(runtime); |
922 | return result; |
923 | } |
924 | |
925 | static int snd_pcm_sw_params(struct snd_pcm_substream *substream, |
926 | struct snd_pcm_sw_params *params) |
927 | { |
928 | struct snd_pcm_runtime *runtime; |
929 | int err; |
930 | |
931 | if (PCM_RUNTIME_CHECK(substream)) |
932 | return -ENXIO; |
933 | runtime = substream->runtime; |
934 | scoped_guard(pcm_stream_lock_irq, substream) { |
935 | if (runtime->state == SNDRV_PCM_STATE_OPEN) |
936 | return -EBADFD; |
937 | } |
938 | |
939 | if (params->tstamp_mode < 0 || |
940 | params->tstamp_mode > SNDRV_PCM_TSTAMP_LAST) |
941 | return -EINVAL; |
942 | if (params->proto >= SNDRV_PROTOCOL_VERSION(2, 0, 12) && |
943 | params->tstamp_type > SNDRV_PCM_TSTAMP_TYPE_LAST) |
944 | return -EINVAL; |
945 | if (params->avail_min == 0) |
946 | return -EINVAL; |
947 | if (params->silence_size >= runtime->boundary) { |
948 | if (params->silence_threshold != 0) |
949 | return -EINVAL; |
950 | } else { |
951 | if (params->silence_size > params->silence_threshold) |
952 | return -EINVAL; |
953 | if (params->silence_threshold > runtime->buffer_size) |
954 | return -EINVAL; |
955 | } |
956 | err = 0; |
957 | scoped_guard(pcm_stream_lock_irq, substream) { |
958 | runtime->tstamp_mode = params->tstamp_mode; |
959 | if (params->proto >= SNDRV_PROTOCOL_VERSION(2, 0, 12)) |
960 | runtime->tstamp_type = params->tstamp_type; |
961 | runtime->period_step = params->period_step; |
962 | runtime->control->avail_min = params->avail_min; |
963 | runtime->start_threshold = params->start_threshold; |
964 | runtime->stop_threshold = params->stop_threshold; |
965 | runtime->silence_threshold = params->silence_threshold; |
966 | runtime->silence_size = params->silence_size; |
967 | params->boundary = runtime->boundary; |
968 | if (snd_pcm_running(substream)) { |
969 | if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && |
970 | runtime->silence_size > 0) |
971 | snd_pcm_playback_silence(substream, ULONG_MAX); |
972 | err = snd_pcm_update_state(substream, runtime); |
973 | } |
974 | } |
975 | return err; |
976 | } |
977 | |
978 | static int snd_pcm_sw_params_user(struct snd_pcm_substream *substream, |
979 | struct snd_pcm_sw_params __user * _params) |
980 | { |
981 | struct snd_pcm_sw_params params; |
982 | int err; |
983 | if (copy_from_user(to: ¶ms, from: _params, n: sizeof(params))) |
984 | return -EFAULT; |
985 | err = snd_pcm_sw_params(substream, params: ¶ms); |
986 | if (copy_to_user(to: _params, from: ¶ms, n: sizeof(params))) |
987 | return -EFAULT; |
988 | return err; |
989 | } |
990 | |
991 | static inline snd_pcm_uframes_t |
992 | snd_pcm_calc_delay(struct snd_pcm_substream *substream) |
993 | { |
994 | snd_pcm_uframes_t delay; |
995 | |
996 | if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) |
997 | delay = snd_pcm_playback_hw_avail(runtime: substream->runtime); |
998 | else |
999 | delay = snd_pcm_capture_avail(runtime: substream->runtime); |
1000 | return delay + substream->runtime->delay; |
1001 | } |
1002 | |
1003 | int snd_pcm_status64(struct snd_pcm_substream *substream, |
1004 | struct snd_pcm_status64 *status) |
1005 | { |
1006 | struct snd_pcm_runtime *runtime = substream->runtime; |
1007 | |
1008 | guard(pcm_stream_lock_irq)(l: substream); |
1009 | |
1010 | snd_pcm_unpack_audio_tstamp_config(data: status->audio_tstamp_data, |
1011 | config: &runtime->audio_tstamp_config); |
1012 | |
1013 | /* backwards compatible behavior */ |
1014 | if (runtime->audio_tstamp_config.type_requested == |
1015 | SNDRV_PCM_AUDIO_TSTAMP_TYPE_COMPAT) { |
1016 | if (runtime->hw.info & SNDRV_PCM_INFO_HAS_WALL_CLOCK) |
1017 | runtime->audio_tstamp_config.type_requested = |
1018 | SNDRV_PCM_AUDIO_TSTAMP_TYPE_LINK; |
1019 | else |
1020 | runtime->audio_tstamp_config.type_requested = |
1021 | SNDRV_PCM_AUDIO_TSTAMP_TYPE_DEFAULT; |
1022 | runtime->audio_tstamp_report.valid = 0; |
1023 | } else |
1024 | runtime->audio_tstamp_report.valid = 1; |
1025 | |
1026 | status->state = runtime->state; |
1027 | status->suspended_state = runtime->suspended_state; |
1028 | if (status->state == SNDRV_PCM_STATE_OPEN) |
1029 | return 0; |
1030 | status->trigger_tstamp_sec = runtime->trigger_tstamp.tv_sec; |
1031 | status->trigger_tstamp_nsec = runtime->trigger_tstamp.tv_nsec; |
1032 | if (snd_pcm_running(substream)) { |
1033 | snd_pcm_update_hw_ptr(substream); |
1034 | if (runtime->tstamp_mode == SNDRV_PCM_TSTAMP_ENABLE) { |
1035 | status->tstamp_sec = runtime->status->tstamp.tv_sec; |
1036 | status->tstamp_nsec = |
1037 | runtime->status->tstamp.tv_nsec; |
1038 | status->driver_tstamp_sec = |
1039 | runtime->driver_tstamp.tv_sec; |
1040 | status->driver_tstamp_nsec = |
1041 | runtime->driver_tstamp.tv_nsec; |
1042 | status->audio_tstamp_sec = |
1043 | runtime->status->audio_tstamp.tv_sec; |
1044 | status->audio_tstamp_nsec = |
1045 | runtime->status->audio_tstamp.tv_nsec; |
1046 | if (runtime->audio_tstamp_report.valid == 1) |
1047 | /* backwards compatibility, no report provided in COMPAT mode */ |
1048 | snd_pcm_pack_audio_tstamp_report(data: &status->audio_tstamp_data, |
1049 | accuracy: &status->audio_tstamp_accuracy, |
1050 | report: &runtime->audio_tstamp_report); |
1051 | |
1052 | goto _tstamp_end; |
1053 | } |
1054 | } else { |
1055 | /* get tstamp only in fallback mode and only if enabled */ |
1056 | if (runtime->tstamp_mode == SNDRV_PCM_TSTAMP_ENABLE) { |
1057 | struct timespec64 tstamp; |
1058 | |
1059 | snd_pcm_gettime(runtime, tv: &tstamp); |
1060 | status->tstamp_sec = tstamp.tv_sec; |
1061 | status->tstamp_nsec = tstamp.tv_nsec; |
1062 | } |
1063 | } |
1064 | _tstamp_end: |
1065 | status->appl_ptr = runtime->control->appl_ptr; |
1066 | status->hw_ptr = runtime->status->hw_ptr; |
1067 | status->avail = snd_pcm_avail(substream); |
1068 | status->delay = snd_pcm_running(substream) ? |
1069 | snd_pcm_calc_delay(substream) : 0; |
1070 | status->avail_max = runtime->avail_max; |
1071 | status->overrange = runtime->overrange; |
1072 | runtime->avail_max = 0; |
1073 | runtime->overrange = 0; |
1074 | return 0; |
1075 | } |
1076 | |
1077 | static int snd_pcm_status_user64(struct snd_pcm_substream *substream, |
1078 | struct snd_pcm_status64 __user * _status, |
1079 | bool ext) |
1080 | { |
1081 | struct snd_pcm_status64 status; |
1082 | int res; |
1083 | |
1084 | memset(&status, 0, sizeof(status)); |
1085 | /* |
1086 | * with extension, parameters are read/write, |
1087 | * get audio_tstamp_data from user, |
1088 | * ignore rest of status structure |
1089 | */ |
1090 | if (ext && get_user(status.audio_tstamp_data, |
1091 | (u32 __user *)(&_status->audio_tstamp_data))) |
1092 | return -EFAULT; |
1093 | res = snd_pcm_status64(substream, status: &status); |
1094 | if (res < 0) |
1095 | return res; |
1096 | if (copy_to_user(to: _status, from: &status, n: sizeof(status))) |
1097 | return -EFAULT; |
1098 | return 0; |
1099 | } |
1100 | |
1101 | static int snd_pcm_status_user32(struct snd_pcm_substream *substream, |
1102 | struct snd_pcm_status32 __user * _status, |
1103 | bool ext) |
1104 | { |
1105 | struct snd_pcm_status64 status64; |
1106 | struct snd_pcm_status32 status32; |
1107 | int res; |
1108 | |
1109 | memset(&status64, 0, sizeof(status64)); |
1110 | memset(&status32, 0, sizeof(status32)); |
1111 | /* |
1112 | * with extension, parameters are read/write, |
1113 | * get audio_tstamp_data from user, |
1114 | * ignore rest of status structure |
1115 | */ |
1116 | if (ext && get_user(status64.audio_tstamp_data, |
1117 | (u32 __user *)(&_status->audio_tstamp_data))) |
1118 | return -EFAULT; |
1119 | res = snd_pcm_status64(substream, status: &status64); |
1120 | if (res < 0) |
1121 | return res; |
1122 | |
1123 | status32 = (struct snd_pcm_status32) { |
1124 | .state = status64.state, |
1125 | .trigger_tstamp_sec = status64.trigger_tstamp_sec, |
1126 | .trigger_tstamp_nsec = status64.trigger_tstamp_nsec, |
1127 | .tstamp_sec = status64.tstamp_sec, |
1128 | .tstamp_nsec = status64.tstamp_nsec, |
1129 | .appl_ptr = status64.appl_ptr, |
1130 | .hw_ptr = status64.hw_ptr, |
1131 | .delay = status64.delay, |
1132 | .avail = status64.avail, |
1133 | .avail_max = status64.avail_max, |
1134 | .overrange = status64.overrange, |
1135 | .suspended_state = status64.suspended_state, |
1136 | .audio_tstamp_data = status64.audio_tstamp_data, |
1137 | .audio_tstamp_sec = status64.audio_tstamp_sec, |
1138 | .audio_tstamp_nsec = status64.audio_tstamp_nsec, |
1139 | .driver_tstamp_sec = status64.audio_tstamp_sec, |
1140 | .driver_tstamp_nsec = status64.audio_tstamp_nsec, |
1141 | .audio_tstamp_accuracy = status64.audio_tstamp_accuracy, |
1142 | }; |
1143 | |
1144 | if (copy_to_user(to: _status, from: &status32, n: sizeof(status32))) |
1145 | return -EFAULT; |
1146 | |
1147 | return 0; |
1148 | } |
1149 | |
1150 | static int snd_pcm_channel_info(struct snd_pcm_substream *substream, |
1151 | struct snd_pcm_channel_info * info) |
1152 | { |
1153 | struct snd_pcm_runtime *runtime; |
1154 | unsigned int channel; |
1155 | |
1156 | channel = info->channel; |
1157 | runtime = substream->runtime; |
1158 | scoped_guard(pcm_stream_lock_irq, substream) { |
1159 | if (runtime->state == SNDRV_PCM_STATE_OPEN) |
1160 | return -EBADFD; |
1161 | } |
1162 | if (channel >= runtime->channels) |
1163 | return -EINVAL; |
1164 | memset(info, 0, sizeof(*info)); |
1165 | info->channel = channel; |
1166 | return snd_pcm_ops_ioctl(substream, SNDRV_PCM_IOCTL1_CHANNEL_INFO, arg: info); |
1167 | } |
1168 | |
1169 | static int snd_pcm_channel_info_user(struct snd_pcm_substream *substream, |
1170 | struct snd_pcm_channel_info __user * _info) |
1171 | { |
1172 | struct snd_pcm_channel_info info; |
1173 | int res; |
1174 | |
1175 | if (copy_from_user(to: &info, from: _info, n: sizeof(info))) |
1176 | return -EFAULT; |
1177 | res = snd_pcm_channel_info(substream, info: &info); |
1178 | if (res < 0) |
1179 | return res; |
1180 | if (copy_to_user(to: _info, from: &info, n: sizeof(info))) |
1181 | return -EFAULT; |
1182 | return 0; |
1183 | } |
1184 | |
1185 | static void snd_pcm_trigger_tstamp(struct snd_pcm_substream *substream) |
1186 | { |
1187 | struct snd_pcm_runtime *runtime = substream->runtime; |
1188 | if (runtime->trigger_master == NULL) |
1189 | return; |
1190 | if (runtime->trigger_master == substream) { |
1191 | if (!runtime->trigger_tstamp_latched) |
1192 | snd_pcm_gettime(runtime, tv: &runtime->trigger_tstamp); |
1193 | } else { |
1194 | snd_pcm_trigger_tstamp(substream: runtime->trigger_master); |
1195 | runtime->trigger_tstamp = runtime->trigger_master->runtime->trigger_tstamp; |
1196 | } |
1197 | runtime->trigger_master = NULL; |
1198 | } |
1199 | |
1200 | #define ACTION_ARG_IGNORE (__force snd_pcm_state_t)0 |
1201 | |
1202 | struct action_ops { |
1203 | int (*pre_action)(struct snd_pcm_substream *substream, |
1204 | snd_pcm_state_t state); |
1205 | int (*do_action)(struct snd_pcm_substream *substream, |
1206 | snd_pcm_state_t state); |
1207 | void (*undo_action)(struct snd_pcm_substream *substream, |
1208 | snd_pcm_state_t state); |
1209 | void (*post_action)(struct snd_pcm_substream *substream, |
1210 | snd_pcm_state_t state); |
1211 | }; |
1212 | |
1213 | /* |
1214 | * this functions is core for handling of linked stream |
1215 | * Note: the stream state might be changed also on failure |
1216 | * Note2: call with calling stream lock + link lock |
1217 | */ |
1218 | static int snd_pcm_action_group(const struct action_ops *ops, |
1219 | struct snd_pcm_substream *substream, |
1220 | snd_pcm_state_t state, |
1221 | bool stream_lock) |
1222 | { |
1223 | struct snd_pcm_substream *s = NULL; |
1224 | struct snd_pcm_substream *s1; |
1225 | int res = 0, depth = 1; |
1226 | |
1227 | snd_pcm_group_for_each_entry(s, substream) { |
1228 | if (s != substream) { |
1229 | if (!stream_lock) |
1230 | mutex_lock_nested(lock: &s->runtime->buffer_mutex, subclass: depth); |
1231 | else if (s->pcm->nonatomic) |
1232 | mutex_lock_nested(lock: &s->self_group.mutex, subclass: depth); |
1233 | else |
1234 | spin_lock_nested(&s->self_group.lock, depth); |
1235 | depth++; |
1236 | } |
1237 | res = ops->pre_action(s, state); |
1238 | if (res < 0) |
1239 | goto _unlock; |
1240 | } |
1241 | snd_pcm_group_for_each_entry(s, substream) { |
1242 | res = ops->do_action(s, state); |
1243 | if (res < 0) { |
1244 | if (ops->undo_action) { |
1245 | snd_pcm_group_for_each_entry(s1, substream) { |
1246 | if (s1 == s) /* failed stream */ |
1247 | break; |
1248 | ops->undo_action(s1, state); |
1249 | } |
1250 | } |
1251 | s = NULL; /* unlock all */ |
1252 | goto _unlock; |
1253 | } |
1254 | } |
1255 | snd_pcm_group_for_each_entry(s, substream) { |
1256 | ops->post_action(s, state); |
1257 | } |
1258 | _unlock: |
1259 | /* unlock streams */ |
1260 | snd_pcm_group_for_each_entry(s1, substream) { |
1261 | if (s1 != substream) { |
1262 | if (!stream_lock) |
1263 | mutex_unlock(lock: &s1->runtime->buffer_mutex); |
1264 | else if (s1->pcm->nonatomic) |
1265 | mutex_unlock(lock: &s1->self_group.mutex); |
1266 | else |
1267 | spin_unlock(lock: &s1->self_group.lock); |
1268 | } |
1269 | if (s1 == s) /* end */ |
1270 | break; |
1271 | } |
1272 | return res; |
1273 | } |
1274 | |
1275 | /* |
1276 | * Note: call with stream lock |
1277 | */ |
1278 | static int snd_pcm_action_single(const struct action_ops *ops, |
1279 | struct snd_pcm_substream *substream, |
1280 | snd_pcm_state_t state) |
1281 | { |
1282 | int res; |
1283 | |
1284 | res = ops->pre_action(substream, state); |
1285 | if (res < 0) |
1286 | return res; |
1287 | res = ops->do_action(substream, state); |
1288 | if (res == 0) |
1289 | ops->post_action(substream, state); |
1290 | else if (ops->undo_action) |
1291 | ops->undo_action(substream, state); |
1292 | return res; |
1293 | } |
1294 | |
1295 | static void snd_pcm_group_assign(struct snd_pcm_substream *substream, |
1296 | struct snd_pcm_group *new_group) |
1297 | { |
1298 | substream->group = new_group; |
1299 | list_move(list: &substream->link_list, head: &new_group->substreams); |
1300 | } |
1301 | |
1302 | /* |
1303 | * Unref and unlock the group, but keep the stream lock; |
1304 | * when the group becomes empty and no longer referred, destroy itself |
1305 | */ |
1306 | static void snd_pcm_group_unref(struct snd_pcm_group *group, |
1307 | struct snd_pcm_substream *substream) |
1308 | { |
1309 | bool do_free; |
1310 | |
1311 | if (!group) |
1312 | return; |
1313 | do_free = refcount_dec_and_test(r: &group->refs); |
1314 | snd_pcm_group_unlock(group, nonatomic: substream->pcm->nonatomic); |
1315 | if (do_free) |
1316 | kfree(objp: group); |
1317 | } |
1318 | |
1319 | /* |
1320 | * Lock the group inside a stream lock and reference it; |
1321 | * return the locked group object, or NULL if not linked |
1322 | */ |
1323 | static struct snd_pcm_group * |
1324 | snd_pcm_stream_group_ref(struct snd_pcm_substream *substream) |
1325 | { |
1326 | bool nonatomic = substream->pcm->nonatomic; |
1327 | struct snd_pcm_group *group; |
1328 | bool trylock; |
1329 | |
1330 | for (;;) { |
1331 | if (!snd_pcm_stream_linked(substream)) |
1332 | return NULL; |
1333 | group = substream->group; |
1334 | /* block freeing the group object */ |
1335 | refcount_inc(r: &group->refs); |
1336 | |
1337 | trylock = nonatomic ? mutex_trylock(lock: &group->mutex) : |
1338 | spin_trylock(lock: &group->lock); |
1339 | if (trylock) |
1340 | break; /* OK */ |
1341 | |
1342 | /* re-lock for avoiding ABBA deadlock */ |
1343 | snd_pcm_stream_unlock(substream); |
1344 | snd_pcm_group_lock(group, nonatomic); |
1345 | snd_pcm_stream_lock(substream); |
1346 | |
1347 | /* check the group again; the above opens a small race window */ |
1348 | if (substream->group == group) |
1349 | break; /* OK */ |
1350 | /* group changed, try again */ |
1351 | snd_pcm_group_unref(group, substream); |
1352 | } |
1353 | return group; |
1354 | } |
1355 | |
1356 | /* |
1357 | * Note: call with stream lock |
1358 | */ |
1359 | static int snd_pcm_action(const struct action_ops *ops, |
1360 | struct snd_pcm_substream *substream, |
1361 | snd_pcm_state_t state) |
1362 | { |
1363 | struct snd_pcm_group *group; |
1364 | int res; |
1365 | |
1366 | group = snd_pcm_stream_group_ref(substream); |
1367 | if (group) |
1368 | res = snd_pcm_action_group(ops, substream, state, stream_lock: true); |
1369 | else |
1370 | res = snd_pcm_action_single(ops, substream, state); |
1371 | snd_pcm_group_unref(group, substream); |
1372 | return res; |
1373 | } |
1374 | |
1375 | /* |
1376 | * Note: don't use any locks before |
1377 | */ |
1378 | static int snd_pcm_action_lock_irq(const struct action_ops *ops, |
1379 | struct snd_pcm_substream *substream, |
1380 | snd_pcm_state_t state) |
1381 | { |
1382 | guard(pcm_stream_lock_irq)(l: substream); |
1383 | return snd_pcm_action(ops, substream, state); |
1384 | } |
1385 | |
1386 | /* |
1387 | */ |
1388 | static int snd_pcm_action_nonatomic(const struct action_ops *ops, |
1389 | struct snd_pcm_substream *substream, |
1390 | snd_pcm_state_t state) |
1391 | { |
1392 | int res; |
1393 | |
1394 | /* Guarantee the group members won't change during non-atomic action */ |
1395 | guard(rwsem_read)(T: &snd_pcm_link_rwsem); |
1396 | res = snd_pcm_buffer_access_lock(runtime: substream->runtime); |
1397 | if (res < 0) |
1398 | return res; |
1399 | if (snd_pcm_stream_linked(substream)) |
1400 | res = snd_pcm_action_group(ops, substream, state, stream_lock: false); |
1401 | else |
1402 | res = snd_pcm_action_single(ops, substream, state); |
1403 | snd_pcm_buffer_access_unlock(runtime: substream->runtime); |
1404 | return res; |
1405 | } |
1406 | |
1407 | /* |
1408 | * start callbacks |
1409 | */ |
1410 | static int snd_pcm_pre_start(struct snd_pcm_substream *substream, |
1411 | snd_pcm_state_t state) |
1412 | { |
1413 | struct snd_pcm_runtime *runtime = substream->runtime; |
1414 | if (runtime->state != SNDRV_PCM_STATE_PREPARED) |
1415 | return -EBADFD; |
1416 | if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && |
1417 | !snd_pcm_playback_data(substream)) |
1418 | return -EPIPE; |
1419 | runtime->trigger_tstamp_latched = false; |
1420 | runtime->trigger_master = substream; |
1421 | return 0; |
1422 | } |
1423 | |
1424 | static int snd_pcm_do_start(struct snd_pcm_substream *substream, |
1425 | snd_pcm_state_t state) |
1426 | { |
1427 | int err; |
1428 | |
1429 | if (substream->runtime->trigger_master != substream) |
1430 | return 0; |
1431 | err = substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_START); |
1432 | /* XRUN happened during the start */ |
1433 | if (err == -EPIPE) |
1434 | __snd_pcm_set_state(runtime: substream->runtime, SNDRV_PCM_STATE_XRUN); |
1435 | return err; |
1436 | } |
1437 | |
1438 | static void snd_pcm_undo_start(struct snd_pcm_substream *substream, |
1439 | snd_pcm_state_t state) |
1440 | { |
1441 | if (substream->runtime->trigger_master == substream) { |
1442 | substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_STOP); |
1443 | substream->runtime->stop_operating = true; |
1444 | } |
1445 | } |
1446 | |
1447 | static void snd_pcm_post_start(struct snd_pcm_substream *substream, |
1448 | snd_pcm_state_t state) |
1449 | { |
1450 | struct snd_pcm_runtime *runtime = substream->runtime; |
1451 | snd_pcm_trigger_tstamp(substream); |
1452 | runtime->hw_ptr_jiffies = jiffies; |
1453 | runtime->hw_ptr_buffer_jiffies = (runtime->buffer_size * HZ) / |
1454 | runtime->rate; |
1455 | __snd_pcm_set_state(runtime, state); |
1456 | if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && |
1457 | runtime->silence_size > 0) |
1458 | snd_pcm_playback_silence(substream, ULONG_MAX); |
1459 | snd_pcm_timer_notify(substream, event: SNDRV_TIMER_EVENT_MSTART); |
1460 | } |
1461 | |
1462 | static const struct action_ops snd_pcm_action_start = { |
1463 | .pre_action = snd_pcm_pre_start, |
1464 | .do_action = snd_pcm_do_start, |
1465 | .undo_action = snd_pcm_undo_start, |
1466 | .post_action = snd_pcm_post_start |
1467 | }; |
1468 | |
1469 | /** |
1470 | * snd_pcm_start - start all linked streams |
1471 | * @substream: the PCM substream instance |
1472 | * |
1473 | * Return: Zero if successful, or a negative error code. |
1474 | * The stream lock must be acquired before calling this function. |
1475 | */ |
1476 | int snd_pcm_start(struct snd_pcm_substream *substream) |
1477 | { |
1478 | return snd_pcm_action(ops: &snd_pcm_action_start, substream, |
1479 | SNDRV_PCM_STATE_RUNNING); |
1480 | } |
1481 | |
1482 | /* take the stream lock and start the streams */ |
1483 | static int snd_pcm_start_lock_irq(struct snd_pcm_substream *substream) |
1484 | { |
1485 | return snd_pcm_action_lock_irq(ops: &snd_pcm_action_start, substream, |
1486 | SNDRV_PCM_STATE_RUNNING); |
1487 | } |
1488 | |
1489 | /* |
1490 | * stop callbacks |
1491 | */ |
1492 | static int snd_pcm_pre_stop(struct snd_pcm_substream *substream, |
1493 | snd_pcm_state_t state) |
1494 | { |
1495 | struct snd_pcm_runtime *runtime = substream->runtime; |
1496 | if (runtime->state == SNDRV_PCM_STATE_OPEN) |
1497 | return -EBADFD; |
1498 | runtime->trigger_master = substream; |
1499 | return 0; |
1500 | } |
1501 | |
1502 | static int snd_pcm_do_stop(struct snd_pcm_substream *substream, |
1503 | snd_pcm_state_t state) |
1504 | { |
1505 | if (substream->runtime->trigger_master == substream && |
1506 | snd_pcm_running(substream)) { |
1507 | substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_STOP); |
1508 | substream->runtime->stop_operating = true; |
1509 | } |
1510 | return 0; /* unconditionally stop all substreams */ |
1511 | } |
1512 | |
1513 | static void snd_pcm_post_stop(struct snd_pcm_substream *substream, |
1514 | snd_pcm_state_t state) |
1515 | { |
1516 | struct snd_pcm_runtime *runtime = substream->runtime; |
1517 | if (runtime->state != state) { |
1518 | snd_pcm_trigger_tstamp(substream); |
1519 | __snd_pcm_set_state(runtime, state); |
1520 | snd_pcm_timer_notify(substream, event: SNDRV_TIMER_EVENT_MSTOP); |
1521 | } |
1522 | wake_up(&runtime->sleep); |
1523 | wake_up(&runtime->tsleep); |
1524 | } |
1525 | |
1526 | static const struct action_ops snd_pcm_action_stop = { |
1527 | .pre_action = snd_pcm_pre_stop, |
1528 | .do_action = snd_pcm_do_stop, |
1529 | .post_action = snd_pcm_post_stop |
1530 | }; |
1531 | |
1532 | /** |
1533 | * snd_pcm_stop - try to stop all running streams in the substream group |
1534 | * @substream: the PCM substream instance |
1535 | * @state: PCM state after stopping the stream |
1536 | * |
1537 | * The state of each stream is then changed to the given state unconditionally. |
1538 | * |
1539 | * Return: Zero if successful, or a negative error code. |
1540 | */ |
1541 | int snd_pcm_stop(struct snd_pcm_substream *substream, snd_pcm_state_t state) |
1542 | { |
1543 | return snd_pcm_action(ops: &snd_pcm_action_stop, substream, state); |
1544 | } |
1545 | EXPORT_SYMBOL(snd_pcm_stop); |
1546 | |
1547 | /** |
1548 | * snd_pcm_drain_done - stop the DMA only when the given stream is playback |
1549 | * @substream: the PCM substream |
1550 | * |
1551 | * After stopping, the state is changed to SETUP. |
1552 | * Unlike snd_pcm_stop(), this affects only the given stream. |
1553 | * |
1554 | * Return: Zero if successful, or a negative error code. |
1555 | */ |
1556 | int snd_pcm_drain_done(struct snd_pcm_substream *substream) |
1557 | { |
1558 | return snd_pcm_action_single(ops: &snd_pcm_action_stop, substream, |
1559 | SNDRV_PCM_STATE_SETUP); |
1560 | } |
1561 | |
1562 | /** |
1563 | * snd_pcm_stop_xrun - stop the running streams as XRUN |
1564 | * @substream: the PCM substream instance |
1565 | * |
1566 | * This stops the given running substream (and all linked substreams) as XRUN. |
1567 | * Unlike snd_pcm_stop(), this function takes the substream lock by itself. |
1568 | * |
1569 | * Return: Zero if successful, or a negative error code. |
1570 | */ |
1571 | int snd_pcm_stop_xrun(struct snd_pcm_substream *substream) |
1572 | { |
1573 | guard(pcm_stream_lock_irqsave)(l: substream); |
1574 | if (substream->runtime && snd_pcm_running(substream)) |
1575 | __snd_pcm_xrun(substream); |
1576 | return 0; |
1577 | } |
1578 | EXPORT_SYMBOL_GPL(snd_pcm_stop_xrun); |
1579 | |
1580 | /* |
1581 | * pause callbacks: pass boolean (to start pause or resume) as state argument |
1582 | */ |
1583 | #define pause_pushed(state) (__force bool)(state) |
1584 | |
1585 | static int snd_pcm_pre_pause(struct snd_pcm_substream *substream, |
1586 | snd_pcm_state_t state) |
1587 | { |
1588 | struct snd_pcm_runtime *runtime = substream->runtime; |
1589 | if (!(runtime->info & SNDRV_PCM_INFO_PAUSE)) |
1590 | return -ENOSYS; |
1591 | if (pause_pushed(state)) { |
1592 | if (runtime->state != SNDRV_PCM_STATE_RUNNING) |
1593 | return -EBADFD; |
1594 | } else if (runtime->state != SNDRV_PCM_STATE_PAUSED) |
1595 | return -EBADFD; |
1596 | runtime->trigger_master = substream; |
1597 | return 0; |
1598 | } |
1599 | |
1600 | static int snd_pcm_do_pause(struct snd_pcm_substream *substream, |
1601 | snd_pcm_state_t state) |
1602 | { |
1603 | if (substream->runtime->trigger_master != substream) |
1604 | return 0; |
1605 | /* The jiffies check in snd_pcm_update_hw_ptr*() is done by |
1606 | * a delta between the current jiffies, this gives a large enough |
1607 | * delta, effectively to skip the check once. |
1608 | */ |
1609 | substream->runtime->hw_ptr_jiffies = jiffies - HZ * 1000; |
1610 | return substream->ops->trigger(substream, |
1611 | pause_pushed(state) ? |
1612 | SNDRV_PCM_TRIGGER_PAUSE_PUSH : |
1613 | SNDRV_PCM_TRIGGER_PAUSE_RELEASE); |
1614 | } |
1615 | |
1616 | static void snd_pcm_undo_pause(struct snd_pcm_substream *substream, |
1617 | snd_pcm_state_t state) |
1618 | { |
1619 | if (substream->runtime->trigger_master == substream) |
1620 | substream->ops->trigger(substream, |
1621 | pause_pushed(state) ? |
1622 | SNDRV_PCM_TRIGGER_PAUSE_RELEASE : |
1623 | SNDRV_PCM_TRIGGER_PAUSE_PUSH); |
1624 | } |
1625 | |
1626 | static void snd_pcm_post_pause(struct snd_pcm_substream *substream, |
1627 | snd_pcm_state_t state) |
1628 | { |
1629 | struct snd_pcm_runtime *runtime = substream->runtime; |
1630 | snd_pcm_trigger_tstamp(substream); |
1631 | if (pause_pushed(state)) { |
1632 | __snd_pcm_set_state(runtime, SNDRV_PCM_STATE_PAUSED); |
1633 | snd_pcm_timer_notify(substream, event: SNDRV_TIMER_EVENT_MPAUSE); |
1634 | wake_up(&runtime->sleep); |
1635 | wake_up(&runtime->tsleep); |
1636 | } else { |
1637 | __snd_pcm_set_state(runtime, SNDRV_PCM_STATE_RUNNING); |
1638 | snd_pcm_timer_notify(substream, event: SNDRV_TIMER_EVENT_MCONTINUE); |
1639 | } |
1640 | } |
1641 | |
1642 | static const struct action_ops snd_pcm_action_pause = { |
1643 | .pre_action = snd_pcm_pre_pause, |
1644 | .do_action = snd_pcm_do_pause, |
1645 | .undo_action = snd_pcm_undo_pause, |
1646 | .post_action = snd_pcm_post_pause |
1647 | }; |
1648 | |
1649 | /* |
1650 | * Push/release the pause for all linked streams. |
1651 | */ |
1652 | static int snd_pcm_pause(struct snd_pcm_substream *substream, bool push) |
1653 | { |
1654 | return snd_pcm_action(ops: &snd_pcm_action_pause, substream, |
1655 | state: (__force snd_pcm_state_t)push); |
1656 | } |
1657 | |
1658 | static int snd_pcm_pause_lock_irq(struct snd_pcm_substream *substream, |
1659 | bool push) |
1660 | { |
1661 | return snd_pcm_action_lock_irq(ops: &snd_pcm_action_pause, substream, |
1662 | state: (__force snd_pcm_state_t)push); |
1663 | } |
1664 | |
1665 | #ifdef CONFIG_PM |
1666 | /* suspend callback: state argument ignored */ |
1667 | |
1668 | static int snd_pcm_pre_suspend(struct snd_pcm_substream *substream, |
1669 | snd_pcm_state_t state) |
1670 | { |
1671 | struct snd_pcm_runtime *runtime = substream->runtime; |
1672 | switch (runtime->state) { |
1673 | case SNDRV_PCM_STATE_SUSPENDED: |
1674 | return -EBUSY; |
1675 | /* unresumable PCM state; return -EBUSY for skipping suspend */ |
1676 | case SNDRV_PCM_STATE_OPEN: |
1677 | case SNDRV_PCM_STATE_SETUP: |
1678 | case SNDRV_PCM_STATE_DISCONNECTED: |
1679 | return -EBUSY; |
1680 | } |
1681 | runtime->trigger_master = substream; |
1682 | return 0; |
1683 | } |
1684 | |
1685 | static int snd_pcm_do_suspend(struct snd_pcm_substream *substream, |
1686 | snd_pcm_state_t state) |
1687 | { |
1688 | struct snd_pcm_runtime *runtime = substream->runtime; |
1689 | if (runtime->trigger_master != substream) |
1690 | return 0; |
1691 | if (! snd_pcm_running(substream)) |
1692 | return 0; |
1693 | substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_SUSPEND); |
1694 | runtime->stop_operating = true; |
1695 | return 0; /* suspend unconditionally */ |
1696 | } |
1697 | |
1698 | static void snd_pcm_post_suspend(struct snd_pcm_substream *substream, |
1699 | snd_pcm_state_t state) |
1700 | { |
1701 | struct snd_pcm_runtime *runtime = substream->runtime; |
1702 | snd_pcm_trigger_tstamp(substream); |
1703 | runtime->suspended_state = runtime->state; |
1704 | runtime->status->suspended_state = runtime->suspended_state; |
1705 | __snd_pcm_set_state(runtime, SNDRV_PCM_STATE_SUSPENDED); |
1706 | snd_pcm_timer_notify(substream, event: SNDRV_TIMER_EVENT_MSUSPEND); |
1707 | wake_up(&runtime->sleep); |
1708 | wake_up(&runtime->tsleep); |
1709 | } |
1710 | |
1711 | static const struct action_ops snd_pcm_action_suspend = { |
1712 | .pre_action = snd_pcm_pre_suspend, |
1713 | .do_action = snd_pcm_do_suspend, |
1714 | .post_action = snd_pcm_post_suspend |
1715 | }; |
1716 | |
1717 | /* |
1718 | * snd_pcm_suspend - trigger SUSPEND to all linked streams |
1719 | * @substream: the PCM substream |
1720 | * |
1721 | * After this call, all streams are changed to SUSPENDED state. |
1722 | * |
1723 | * Return: Zero if successful, or a negative error code. |
1724 | */ |
1725 | static int snd_pcm_suspend(struct snd_pcm_substream *substream) |
1726 | { |
1727 | guard(pcm_stream_lock_irqsave)(l: substream); |
1728 | return snd_pcm_action(ops: &snd_pcm_action_suspend, substream, |
1729 | ACTION_ARG_IGNORE); |
1730 | } |
1731 | |
1732 | /** |
1733 | * snd_pcm_suspend_all - trigger SUSPEND to all substreams in the given pcm |
1734 | * @pcm: the PCM instance |
1735 | * |
1736 | * After this call, all streams are changed to SUSPENDED state. |
1737 | * |
1738 | * Return: Zero if successful (or @pcm is %NULL), or a negative error code. |
1739 | */ |
1740 | int snd_pcm_suspend_all(struct snd_pcm *pcm) |
1741 | { |
1742 | struct snd_pcm_substream *substream; |
1743 | int stream, err = 0; |
1744 | |
1745 | if (! pcm) |
1746 | return 0; |
1747 | |
1748 | for_each_pcm_substream(pcm, stream, substream) { |
1749 | /* FIXME: the open/close code should lock this as well */ |
1750 | if (!substream->runtime) |
1751 | continue; |
1752 | |
1753 | /* |
1754 | * Skip BE dai link PCM's that are internal and may |
1755 | * not have their substream ops set. |
1756 | */ |
1757 | if (!substream->ops) |
1758 | continue; |
1759 | |
1760 | err = snd_pcm_suspend(substream); |
1761 | if (err < 0 && err != -EBUSY) |
1762 | return err; |
1763 | } |
1764 | |
1765 | for_each_pcm_substream(pcm, stream, substream) |
1766 | snd_pcm_sync_stop(substream, sync_irq: false); |
1767 | |
1768 | return 0; |
1769 | } |
1770 | EXPORT_SYMBOL(snd_pcm_suspend_all); |
1771 | |
1772 | /* resume callbacks: state argument ignored */ |
1773 | |
1774 | static int snd_pcm_pre_resume(struct snd_pcm_substream *substream, |
1775 | snd_pcm_state_t state) |
1776 | { |
1777 | struct snd_pcm_runtime *runtime = substream->runtime; |
1778 | if (!(runtime->info & SNDRV_PCM_INFO_RESUME)) |
1779 | return -ENOSYS; |
1780 | runtime->trigger_master = substream; |
1781 | return 0; |
1782 | } |
1783 | |
1784 | static int snd_pcm_do_resume(struct snd_pcm_substream *substream, |
1785 | snd_pcm_state_t state) |
1786 | { |
1787 | struct snd_pcm_runtime *runtime = substream->runtime; |
1788 | if (runtime->trigger_master != substream) |
1789 | return 0; |
1790 | /* DMA not running previously? */ |
1791 | if (runtime->suspended_state != SNDRV_PCM_STATE_RUNNING && |
1792 | (runtime->suspended_state != SNDRV_PCM_STATE_DRAINING || |
1793 | substream->stream != SNDRV_PCM_STREAM_PLAYBACK)) |
1794 | return 0; |
1795 | return substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_RESUME); |
1796 | } |
1797 | |
1798 | static void snd_pcm_undo_resume(struct snd_pcm_substream *substream, |
1799 | snd_pcm_state_t state) |
1800 | { |
1801 | if (substream->runtime->trigger_master == substream && |
1802 | snd_pcm_running(substream)) |
1803 | substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_SUSPEND); |
1804 | } |
1805 | |
1806 | static void snd_pcm_post_resume(struct snd_pcm_substream *substream, |
1807 | snd_pcm_state_t state) |
1808 | { |
1809 | struct snd_pcm_runtime *runtime = substream->runtime; |
1810 | snd_pcm_trigger_tstamp(substream); |
1811 | __snd_pcm_set_state(runtime, state: runtime->suspended_state); |
1812 | snd_pcm_timer_notify(substream, event: SNDRV_TIMER_EVENT_MRESUME); |
1813 | } |
1814 | |
1815 | static const struct action_ops snd_pcm_action_resume = { |
1816 | .pre_action = snd_pcm_pre_resume, |
1817 | .do_action = snd_pcm_do_resume, |
1818 | .undo_action = snd_pcm_undo_resume, |
1819 | .post_action = snd_pcm_post_resume |
1820 | }; |
1821 | |
1822 | static int snd_pcm_resume(struct snd_pcm_substream *substream) |
1823 | { |
1824 | return snd_pcm_action_lock_irq(ops: &snd_pcm_action_resume, substream, |
1825 | ACTION_ARG_IGNORE); |
1826 | } |
1827 | |
1828 | #else |
1829 | |
1830 | static int snd_pcm_resume(struct snd_pcm_substream *substream) |
1831 | { |
1832 | return -ENOSYS; |
1833 | } |
1834 | |
1835 | #endif /* CONFIG_PM */ |
1836 | |
1837 | /* |
1838 | * xrun ioctl |
1839 | * |
1840 | * Change the RUNNING stream(s) to XRUN state. |
1841 | */ |
1842 | static int snd_pcm_xrun(struct snd_pcm_substream *substream) |
1843 | { |
1844 | struct snd_pcm_runtime *runtime = substream->runtime; |
1845 | |
1846 | guard(pcm_stream_lock_irq)(l: substream); |
1847 | switch (runtime->state) { |
1848 | case SNDRV_PCM_STATE_XRUN: |
1849 | return 0; /* already there */ |
1850 | case SNDRV_PCM_STATE_RUNNING: |
1851 | __snd_pcm_xrun(substream); |
1852 | return 0; |
1853 | default: |
1854 | return -EBADFD; |
1855 | } |
1856 | } |
1857 | |
1858 | /* |
1859 | * reset ioctl |
1860 | */ |
1861 | /* reset callbacks: state argument ignored */ |
1862 | static int snd_pcm_pre_reset(struct snd_pcm_substream *substream, |
1863 | snd_pcm_state_t state) |
1864 | { |
1865 | struct snd_pcm_runtime *runtime = substream->runtime; |
1866 | switch (runtime->state) { |
1867 | case SNDRV_PCM_STATE_RUNNING: |
1868 | case SNDRV_PCM_STATE_PREPARED: |
1869 | case SNDRV_PCM_STATE_PAUSED: |
1870 | case SNDRV_PCM_STATE_SUSPENDED: |
1871 | return 0; |
1872 | default: |
1873 | return -EBADFD; |
1874 | } |
1875 | } |
1876 | |
1877 | static int snd_pcm_do_reset(struct snd_pcm_substream *substream, |
1878 | snd_pcm_state_t state) |
1879 | { |
1880 | struct snd_pcm_runtime *runtime = substream->runtime; |
1881 | int err = snd_pcm_ops_ioctl(substream, SNDRV_PCM_IOCTL1_RESET, NULL); |
1882 | if (err < 0) |
1883 | return err; |
1884 | guard(pcm_stream_lock_irq)(l: substream); |
1885 | runtime->hw_ptr_base = 0; |
1886 | runtime->hw_ptr_interrupt = runtime->status->hw_ptr - |
1887 | runtime->status->hw_ptr % runtime->period_size; |
1888 | runtime->silence_start = runtime->status->hw_ptr; |
1889 | runtime->silence_filled = 0; |
1890 | return 0; |
1891 | } |
1892 | |
1893 | static void snd_pcm_post_reset(struct snd_pcm_substream *substream, |
1894 | snd_pcm_state_t state) |
1895 | { |
1896 | struct snd_pcm_runtime *runtime = substream->runtime; |
1897 | guard(pcm_stream_lock_irq)(l: substream); |
1898 | runtime->control->appl_ptr = runtime->status->hw_ptr; |
1899 | if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && |
1900 | runtime->silence_size > 0) |
1901 | snd_pcm_playback_silence(substream, ULONG_MAX); |
1902 | } |
1903 | |
1904 | static const struct action_ops snd_pcm_action_reset = { |
1905 | .pre_action = snd_pcm_pre_reset, |
1906 | .do_action = snd_pcm_do_reset, |
1907 | .post_action = snd_pcm_post_reset |
1908 | }; |
1909 | |
1910 | static int snd_pcm_reset(struct snd_pcm_substream *substream) |
1911 | { |
1912 | return snd_pcm_action_nonatomic(ops: &snd_pcm_action_reset, substream, |
1913 | ACTION_ARG_IGNORE); |
1914 | } |
1915 | |
1916 | /* |
1917 | * prepare ioctl |
1918 | */ |
1919 | /* pass f_flags as state argument */ |
1920 | static int snd_pcm_pre_prepare(struct snd_pcm_substream *substream, |
1921 | snd_pcm_state_t state) |
1922 | { |
1923 | struct snd_pcm_runtime *runtime = substream->runtime; |
1924 | int f_flags = (__force int)state; |
1925 | |
1926 | if (runtime->state == SNDRV_PCM_STATE_OPEN || |
1927 | runtime->state == SNDRV_PCM_STATE_DISCONNECTED) |
1928 | return -EBADFD; |
1929 | if (snd_pcm_running(substream)) |
1930 | return -EBUSY; |
1931 | substream->f_flags = f_flags; |
1932 | return 0; |
1933 | } |
1934 | |
1935 | static int snd_pcm_do_prepare(struct snd_pcm_substream *substream, |
1936 | snd_pcm_state_t state) |
1937 | { |
1938 | int err; |
1939 | snd_pcm_sync_stop(substream, sync_irq: true); |
1940 | err = substream->ops->prepare(substream); |
1941 | if (err < 0) |
1942 | return err; |
1943 | return snd_pcm_do_reset(substream, state); |
1944 | } |
1945 | |
1946 | static void snd_pcm_post_prepare(struct snd_pcm_substream *substream, |
1947 | snd_pcm_state_t state) |
1948 | { |
1949 | struct snd_pcm_runtime *runtime = substream->runtime; |
1950 | runtime->control->appl_ptr = runtime->status->hw_ptr; |
1951 | snd_pcm_set_state(substream, SNDRV_PCM_STATE_PREPARED); |
1952 | } |
1953 | |
1954 | static const struct action_ops snd_pcm_action_prepare = { |
1955 | .pre_action = snd_pcm_pre_prepare, |
1956 | .do_action = snd_pcm_do_prepare, |
1957 | .post_action = snd_pcm_post_prepare |
1958 | }; |
1959 | |
1960 | /** |
1961 | * snd_pcm_prepare - prepare the PCM substream to be triggerable |
1962 | * @substream: the PCM substream instance |
1963 | * @file: file to refer f_flags |
1964 | * |
1965 | * Return: Zero if successful, or a negative error code. |
1966 | */ |
1967 | static int snd_pcm_prepare(struct snd_pcm_substream *substream, |
1968 | struct file *file) |
1969 | { |
1970 | int f_flags; |
1971 | |
1972 | if (file) |
1973 | f_flags = file->f_flags; |
1974 | else |
1975 | f_flags = substream->f_flags; |
1976 | |
1977 | scoped_guard(pcm_stream_lock_irq, substream) { |
1978 | switch (substream->runtime->state) { |
1979 | case SNDRV_PCM_STATE_PAUSED: |
1980 | snd_pcm_pause(substream, push: false); |
1981 | fallthrough; |
1982 | case SNDRV_PCM_STATE_SUSPENDED: |
1983 | snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP); |
1984 | break; |
1985 | } |
1986 | } |
1987 | |
1988 | return snd_pcm_action_nonatomic(ops: &snd_pcm_action_prepare, |
1989 | substream, |
1990 | state: (__force snd_pcm_state_t)f_flags); |
1991 | } |
1992 | |
1993 | /* |
1994 | * drain ioctl |
1995 | */ |
1996 | |
1997 | /* drain init callbacks: state argument ignored */ |
1998 | static int snd_pcm_pre_drain_init(struct snd_pcm_substream *substream, |
1999 | snd_pcm_state_t state) |
2000 | { |
2001 | struct snd_pcm_runtime *runtime = substream->runtime; |
2002 | switch (runtime->state) { |
2003 | case SNDRV_PCM_STATE_OPEN: |
2004 | case SNDRV_PCM_STATE_DISCONNECTED: |
2005 | case SNDRV_PCM_STATE_SUSPENDED: |
2006 | return -EBADFD; |
2007 | } |
2008 | runtime->trigger_master = substream; |
2009 | return 0; |
2010 | } |
2011 | |
2012 | static int snd_pcm_do_drain_init(struct snd_pcm_substream *substream, |
2013 | snd_pcm_state_t state) |
2014 | { |
2015 | struct snd_pcm_runtime *runtime = substream->runtime; |
2016 | if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) { |
2017 | switch (runtime->state) { |
2018 | case SNDRV_PCM_STATE_PREPARED: |
2019 | /* start playback stream if possible */ |
2020 | if (! snd_pcm_playback_empty(substream)) { |
2021 | snd_pcm_do_start(substream, SNDRV_PCM_STATE_DRAINING); |
2022 | snd_pcm_post_start(substream, SNDRV_PCM_STATE_DRAINING); |
2023 | } else { |
2024 | __snd_pcm_set_state(runtime, SNDRV_PCM_STATE_SETUP); |
2025 | } |
2026 | break; |
2027 | case SNDRV_PCM_STATE_RUNNING: |
2028 | __snd_pcm_set_state(runtime, SNDRV_PCM_STATE_DRAINING); |
2029 | break; |
2030 | case SNDRV_PCM_STATE_XRUN: |
2031 | __snd_pcm_set_state(runtime, SNDRV_PCM_STATE_SETUP); |
2032 | break; |
2033 | default: |
2034 | break; |
2035 | } |
2036 | } else { |
2037 | /* stop running stream */ |
2038 | if (runtime->state == SNDRV_PCM_STATE_RUNNING) { |
2039 | snd_pcm_state_t new_state; |
2040 | |
2041 | new_state = snd_pcm_capture_avail(runtime) > 0 ? |
2042 | SNDRV_PCM_STATE_DRAINING : SNDRV_PCM_STATE_SETUP; |
2043 | snd_pcm_do_stop(substream, state: new_state); |
2044 | snd_pcm_post_stop(substream, state: new_state); |
2045 | } |
2046 | } |
2047 | |
2048 | if (runtime->state == SNDRV_PCM_STATE_DRAINING && |
2049 | runtime->trigger_master == substream && |
2050 | (runtime->hw.info & SNDRV_PCM_INFO_DRAIN_TRIGGER)) |
2051 | return substream->ops->trigger(substream, |
2052 | SNDRV_PCM_TRIGGER_DRAIN); |
2053 | |
2054 | return 0; |
2055 | } |
2056 | |
2057 | static void snd_pcm_post_drain_init(struct snd_pcm_substream *substream, |
2058 | snd_pcm_state_t state) |
2059 | { |
2060 | } |
2061 | |
2062 | static const struct action_ops snd_pcm_action_drain_init = { |
2063 | .pre_action = snd_pcm_pre_drain_init, |
2064 | .do_action = snd_pcm_do_drain_init, |
2065 | .post_action = snd_pcm_post_drain_init |
2066 | }; |
2067 | |
2068 | /* |
2069 | * Drain the stream(s). |
2070 | * When the substream is linked, sync until the draining of all playback streams |
2071 | * is finished. |
2072 | * After this call, all streams are supposed to be either SETUP or DRAINING |
2073 | * (capture only) state. |
2074 | */ |
2075 | static int snd_pcm_drain(struct snd_pcm_substream *substream, |
2076 | struct file *file) |
2077 | { |
2078 | struct snd_card *card; |
2079 | struct snd_pcm_runtime *runtime; |
2080 | struct snd_pcm_substream *s; |
2081 | struct snd_pcm_group *group; |
2082 | wait_queue_entry_t wait; |
2083 | int result = 0; |
2084 | int nonblock = 0; |
2085 | |
2086 | card = substream->pcm->card; |
2087 | runtime = substream->runtime; |
2088 | |
2089 | if (runtime->state == SNDRV_PCM_STATE_OPEN) |
2090 | return -EBADFD; |
2091 | |
2092 | if (file) { |
2093 | if (file->f_flags & O_NONBLOCK) |
2094 | nonblock = 1; |
2095 | } else if (substream->f_flags & O_NONBLOCK) |
2096 | nonblock = 1; |
2097 | |
2098 | snd_pcm_stream_lock_irq(substream); |
2099 | /* resume pause */ |
2100 | if (runtime->state == SNDRV_PCM_STATE_PAUSED) |
2101 | snd_pcm_pause(substream, push: false); |
2102 | |
2103 | /* pre-start/stop - all running streams are changed to DRAINING state */ |
2104 | result = snd_pcm_action(ops: &snd_pcm_action_drain_init, substream, |
2105 | ACTION_ARG_IGNORE); |
2106 | if (result < 0) |
2107 | goto unlock; |
2108 | /* in non-blocking, we don't wait in ioctl but let caller poll */ |
2109 | if (nonblock) { |
2110 | result = -EAGAIN; |
2111 | goto unlock; |
2112 | } |
2113 | |
2114 | for (;;) { |
2115 | long tout; |
2116 | struct snd_pcm_runtime *to_check; |
2117 | if (signal_pending(current)) { |
2118 | result = -ERESTARTSYS; |
2119 | break; |
2120 | } |
2121 | /* find a substream to drain */ |
2122 | to_check = NULL; |
2123 | group = snd_pcm_stream_group_ref(substream); |
2124 | snd_pcm_group_for_each_entry(s, substream) { |
2125 | if (s->stream != SNDRV_PCM_STREAM_PLAYBACK) |
2126 | continue; |
2127 | runtime = s->runtime; |
2128 | if (runtime->state == SNDRV_PCM_STATE_DRAINING) { |
2129 | to_check = runtime; |
2130 | break; |
2131 | } |
2132 | } |
2133 | snd_pcm_group_unref(group, substream); |
2134 | if (!to_check) |
2135 | break; /* all drained */ |
2136 | init_waitqueue_entry(wq_entry: &wait, current); |
2137 | set_current_state(TASK_INTERRUPTIBLE); |
2138 | add_wait_queue(wq_head: &to_check->sleep, wq_entry: &wait); |
2139 | snd_pcm_stream_unlock_irq(substream); |
2140 | if (runtime->no_period_wakeup) |
2141 | tout = MAX_SCHEDULE_TIMEOUT; |
2142 | else { |
2143 | tout = 100; |
2144 | if (runtime->rate) { |
2145 | long t = runtime->buffer_size * 1100 / runtime->rate; |
2146 | tout = max(t, tout); |
2147 | } |
2148 | tout = msecs_to_jiffies(m: tout); |
2149 | } |
2150 | tout = schedule_timeout(timeout: tout); |
2151 | |
2152 | snd_pcm_stream_lock_irq(substream); |
2153 | group = snd_pcm_stream_group_ref(substream); |
2154 | snd_pcm_group_for_each_entry(s, substream) { |
2155 | if (s->runtime == to_check) { |
2156 | remove_wait_queue(wq_head: &to_check->sleep, wq_entry: &wait); |
2157 | break; |
2158 | } |
2159 | } |
2160 | snd_pcm_group_unref(group, substream); |
2161 | |
2162 | if (card->shutdown) { |
2163 | result = -ENODEV; |
2164 | break; |
2165 | } |
2166 | if (tout == 0) { |
2167 | if (substream->runtime->state == SNDRV_PCM_STATE_SUSPENDED) |
2168 | result = -ESTRPIPE; |
2169 | else { |
2170 | dev_dbg(substream->pcm->card->dev, |
2171 | "playback drain timeout (DMA or IRQ trouble?)\n" ); |
2172 | snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP); |
2173 | result = -EIO; |
2174 | } |
2175 | break; |
2176 | } |
2177 | } |
2178 | |
2179 | unlock: |
2180 | snd_pcm_stream_unlock_irq(substream); |
2181 | |
2182 | return result; |
2183 | } |
2184 | |
2185 | /* |
2186 | * drop ioctl |
2187 | * |
2188 | * Immediately put all linked substreams into SETUP state. |
2189 | */ |
2190 | static int snd_pcm_drop(struct snd_pcm_substream *substream) |
2191 | { |
2192 | struct snd_pcm_runtime *runtime; |
2193 | int result = 0; |
2194 | |
2195 | if (PCM_RUNTIME_CHECK(substream)) |
2196 | return -ENXIO; |
2197 | runtime = substream->runtime; |
2198 | |
2199 | if (runtime->state == SNDRV_PCM_STATE_OPEN || |
2200 | runtime->state == SNDRV_PCM_STATE_DISCONNECTED) |
2201 | return -EBADFD; |
2202 | |
2203 | guard(pcm_stream_lock_irq)(l: substream); |
2204 | /* resume pause */ |
2205 | if (runtime->state == SNDRV_PCM_STATE_PAUSED) |
2206 | snd_pcm_pause(substream, push: false); |
2207 | |
2208 | snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP); |
2209 | /* runtime->control->appl_ptr = runtime->status->hw_ptr; */ |
2210 | |
2211 | return result; |
2212 | } |
2213 | |
2214 | |
2215 | static bool is_pcm_file(struct file *file) |
2216 | { |
2217 | struct inode *inode = file_inode(f: file); |
2218 | struct snd_pcm *pcm; |
2219 | unsigned int minor; |
2220 | |
2221 | if (!S_ISCHR(inode->i_mode) || imajor(inode) != snd_major) |
2222 | return false; |
2223 | minor = iminor(inode); |
2224 | pcm = snd_lookup_minor_data(minor, type: SNDRV_DEVICE_TYPE_PCM_PLAYBACK); |
2225 | if (!pcm) |
2226 | pcm = snd_lookup_minor_data(minor, type: SNDRV_DEVICE_TYPE_PCM_CAPTURE); |
2227 | if (!pcm) |
2228 | return false; |
2229 | snd_card_unref(card: pcm->card); |
2230 | return true; |
2231 | } |
2232 | |
2233 | /* |
2234 | * PCM link handling |
2235 | */ |
2236 | static int snd_pcm_link(struct snd_pcm_substream *substream, int fd) |
2237 | { |
2238 | struct snd_pcm_file *pcm_file; |
2239 | struct snd_pcm_substream *substream1; |
2240 | struct snd_pcm_group *group __free(kfree) = NULL; |
2241 | struct snd_pcm_group *target_group; |
2242 | bool nonatomic = substream->pcm->nonatomic; |
2243 | CLASS(fd, f)(fd); |
2244 | |
2245 | if (!f.file) |
2246 | return -EBADFD; |
2247 | if (!is_pcm_file(file: f.file)) |
2248 | return -EBADFD; |
2249 | |
2250 | pcm_file = f.file->private_data; |
2251 | substream1 = pcm_file->substream; |
2252 | |
2253 | if (substream == substream1) |
2254 | return -EINVAL; |
2255 | |
2256 | group = kzalloc(size: sizeof(*group), GFP_KERNEL); |
2257 | if (!group) |
2258 | return -ENOMEM; |
2259 | snd_pcm_group_init(group); |
2260 | |
2261 | guard(rwsem_write)(T: &snd_pcm_link_rwsem); |
2262 | if (substream->runtime->state == SNDRV_PCM_STATE_OPEN || |
2263 | substream->runtime->state != substream1->runtime->state || |
2264 | substream->pcm->nonatomic != substream1->pcm->nonatomic) |
2265 | return -EBADFD; |
2266 | if (snd_pcm_stream_linked(substream: substream1)) |
2267 | return -EALREADY; |
2268 | |
2269 | scoped_guard(pcm_stream_lock_irq, substream) { |
2270 | if (!snd_pcm_stream_linked(substream)) { |
2271 | snd_pcm_group_assign(substream, new_group: group); |
2272 | group = NULL; /* assigned, don't free this one below */ |
2273 | } |
2274 | target_group = substream->group; |
2275 | } |
2276 | |
2277 | snd_pcm_group_lock_irq(group: target_group, nonatomic); |
2278 | snd_pcm_stream_lock_nested(substream: substream1); |
2279 | snd_pcm_group_assign(substream: substream1, new_group: target_group); |
2280 | refcount_inc(r: &target_group->refs); |
2281 | snd_pcm_stream_unlock(substream1); |
2282 | snd_pcm_group_unlock_irq(group: target_group, nonatomic); |
2283 | return 0; |
2284 | } |
2285 | |
2286 | static void relink_to_local(struct snd_pcm_substream *substream) |
2287 | { |
2288 | snd_pcm_stream_lock_nested(substream); |
2289 | snd_pcm_group_assign(substream, new_group: &substream->self_group); |
2290 | snd_pcm_stream_unlock(substream); |
2291 | } |
2292 | |
2293 | static int snd_pcm_unlink(struct snd_pcm_substream *substream) |
2294 | { |
2295 | struct snd_pcm_group *group; |
2296 | bool nonatomic = substream->pcm->nonatomic; |
2297 | bool do_free = false; |
2298 | |
2299 | guard(rwsem_write)(T: &snd_pcm_link_rwsem); |
2300 | |
2301 | if (!snd_pcm_stream_linked(substream)) |
2302 | return -EALREADY; |
2303 | |
2304 | group = substream->group; |
2305 | snd_pcm_group_lock_irq(group, nonatomic); |
2306 | |
2307 | relink_to_local(substream); |
2308 | refcount_dec(r: &group->refs); |
2309 | |
2310 | /* detach the last stream, too */ |
2311 | if (list_is_singular(head: &group->substreams)) { |
2312 | relink_to_local(list_first_entry(&group->substreams, |
2313 | struct snd_pcm_substream, |
2314 | link_list)); |
2315 | do_free = refcount_dec_and_test(r: &group->refs); |
2316 | } |
2317 | |
2318 | snd_pcm_group_unlock_irq(group, nonatomic); |
2319 | if (do_free) |
2320 | kfree(objp: group); |
2321 | return 0; |
2322 | } |
2323 | |
2324 | /* |
2325 | * hw configurator |
2326 | */ |
2327 | static int snd_pcm_hw_rule_mul(struct snd_pcm_hw_params *params, |
2328 | struct snd_pcm_hw_rule *rule) |
2329 | { |
2330 | struct snd_interval t; |
2331 | snd_interval_mul(a: hw_param_interval_c(params, var: rule->deps[0]), |
2332 | b: hw_param_interval_c(params, var: rule->deps[1]), c: &t); |
2333 | return snd_interval_refine(i: hw_param_interval(params, var: rule->var), v: &t); |
2334 | } |
2335 | |
2336 | static int snd_pcm_hw_rule_div(struct snd_pcm_hw_params *params, |
2337 | struct snd_pcm_hw_rule *rule) |
2338 | { |
2339 | struct snd_interval t; |
2340 | snd_interval_div(a: hw_param_interval_c(params, var: rule->deps[0]), |
2341 | b: hw_param_interval_c(params, var: rule->deps[1]), c: &t); |
2342 | return snd_interval_refine(i: hw_param_interval(params, var: rule->var), v: &t); |
2343 | } |
2344 | |
2345 | static int snd_pcm_hw_rule_muldivk(struct snd_pcm_hw_params *params, |
2346 | struct snd_pcm_hw_rule *rule) |
2347 | { |
2348 | struct snd_interval t; |
2349 | snd_interval_muldivk(a: hw_param_interval_c(params, var: rule->deps[0]), |
2350 | b: hw_param_interval_c(params, var: rule->deps[1]), |
2351 | k: (unsigned long) rule->private, c: &t); |
2352 | return snd_interval_refine(i: hw_param_interval(params, var: rule->var), v: &t); |
2353 | } |
2354 | |
2355 | static int snd_pcm_hw_rule_mulkdiv(struct snd_pcm_hw_params *params, |
2356 | struct snd_pcm_hw_rule *rule) |
2357 | { |
2358 | struct snd_interval t; |
2359 | snd_interval_mulkdiv(a: hw_param_interval_c(params, var: rule->deps[0]), |
2360 | k: (unsigned long) rule->private, |
2361 | b: hw_param_interval_c(params, var: rule->deps[1]), c: &t); |
2362 | return snd_interval_refine(i: hw_param_interval(params, var: rule->var), v: &t); |
2363 | } |
2364 | |
2365 | static int snd_pcm_hw_rule_format(struct snd_pcm_hw_params *params, |
2366 | struct snd_pcm_hw_rule *rule) |
2367 | { |
2368 | snd_pcm_format_t k; |
2369 | const struct snd_interval *i = |
2370 | hw_param_interval_c(params, var: rule->deps[0]); |
2371 | struct snd_mask m; |
2372 | struct snd_mask *mask = hw_param_mask(params, SNDRV_PCM_HW_PARAM_FORMAT); |
2373 | snd_mask_any(mask: &m); |
2374 | pcm_for_each_format(k) { |
2375 | int bits; |
2376 | if (!snd_mask_test_format(mask, format: k)) |
2377 | continue; |
2378 | bits = snd_pcm_format_physical_width(format: k); |
2379 | if (bits <= 0) |
2380 | continue; /* ignore invalid formats */ |
2381 | if ((unsigned)bits < i->min || (unsigned)bits > i->max) |
2382 | snd_mask_reset(mask: &m, val: (__force unsigned)k); |
2383 | } |
2384 | return snd_mask_refine(mask, v: &m); |
2385 | } |
2386 | |
2387 | static int snd_pcm_hw_rule_sample_bits(struct snd_pcm_hw_params *params, |
2388 | struct snd_pcm_hw_rule *rule) |
2389 | { |
2390 | struct snd_interval t; |
2391 | snd_pcm_format_t k; |
2392 | |
2393 | t.min = UINT_MAX; |
2394 | t.max = 0; |
2395 | t.openmin = 0; |
2396 | t.openmax = 0; |
2397 | pcm_for_each_format(k) { |
2398 | int bits; |
2399 | if (!snd_mask_test_format(mask: hw_param_mask(params, SNDRV_PCM_HW_PARAM_FORMAT), format: k)) |
2400 | continue; |
2401 | bits = snd_pcm_format_physical_width(format: k); |
2402 | if (bits <= 0) |
2403 | continue; /* ignore invalid formats */ |
2404 | if (t.min > (unsigned)bits) |
2405 | t.min = bits; |
2406 | if (t.max < (unsigned)bits) |
2407 | t.max = bits; |
2408 | } |
2409 | t.integer = 1; |
2410 | return snd_interval_refine(i: hw_param_interval(params, var: rule->var), v: &t); |
2411 | } |
2412 | |
2413 | #if SNDRV_PCM_RATE_5512 != 1 << 0 || SNDRV_PCM_RATE_192000 != 1 << 12 |
2414 | #error "Change this table" |
2415 | #endif |
2416 | |
2417 | static const unsigned int rates[] = { |
2418 | 5512, 8000, 11025, 16000, 22050, 32000, 44100, |
2419 | 48000, 64000, 88200, 96000, 176400, 192000, 352800, 384000 |
2420 | }; |
2421 | |
2422 | const struct snd_pcm_hw_constraint_list snd_pcm_known_rates = { |
2423 | .count = ARRAY_SIZE(rates), |
2424 | .list = rates, |
2425 | }; |
2426 | |
2427 | static int snd_pcm_hw_rule_rate(struct snd_pcm_hw_params *params, |
2428 | struct snd_pcm_hw_rule *rule) |
2429 | { |
2430 | struct snd_pcm_hardware *hw = rule->private; |
2431 | return snd_interval_list(i: hw_param_interval(params, var: rule->var), |
2432 | count: snd_pcm_known_rates.count, |
2433 | list: snd_pcm_known_rates.list, mask: hw->rates); |
2434 | } |
2435 | |
2436 | static int snd_pcm_hw_rule_buffer_bytes_max(struct snd_pcm_hw_params *params, |
2437 | struct snd_pcm_hw_rule *rule) |
2438 | { |
2439 | struct snd_interval t; |
2440 | struct snd_pcm_substream *substream = rule->private; |
2441 | t.min = 0; |
2442 | t.max = substream->buffer_bytes_max; |
2443 | t.openmin = 0; |
2444 | t.openmax = 0; |
2445 | t.integer = 1; |
2446 | return snd_interval_refine(i: hw_param_interval(params, var: rule->var), v: &t); |
2447 | } |
2448 | |
2449 | static int snd_pcm_hw_rule_subformats(struct snd_pcm_hw_params *params, |
2450 | struct snd_pcm_hw_rule *rule) |
2451 | { |
2452 | struct snd_mask *sfmask = hw_param_mask(params, SNDRV_PCM_HW_PARAM_SUBFORMAT); |
2453 | struct snd_mask *fmask = hw_param_mask(params, SNDRV_PCM_HW_PARAM_FORMAT); |
2454 | u32 *subformats = rule->private; |
2455 | snd_pcm_format_t f; |
2456 | struct snd_mask m; |
2457 | |
2458 | snd_mask_none(mask: &m); |
2459 | /* All PCMs support at least the default STD subformat. */ |
2460 | snd_mask_set(mask: &m, val: (__force unsigned)SNDRV_PCM_SUBFORMAT_STD); |
2461 | |
2462 | pcm_for_each_format(f) { |
2463 | if (!snd_mask_test(mask: fmask, val: (__force unsigned)f)) |
2464 | continue; |
2465 | |
2466 | if (f == SNDRV_PCM_FORMAT_S32_LE && *subformats) |
2467 | m.bits[0] |= *subformats; |
2468 | else if (snd_pcm_format_linear(format: f)) |
2469 | snd_mask_set(mask: &m, val: (__force unsigned)SNDRV_PCM_SUBFORMAT_MSBITS_MAX); |
2470 | } |
2471 | |
2472 | return snd_mask_refine(mask: sfmask, v: &m); |
2473 | } |
2474 | |
2475 | static int snd_pcm_hw_constraint_subformats(struct snd_pcm_runtime *runtime, |
2476 | unsigned int cond, u32 *subformats) |
2477 | { |
2478 | return snd_pcm_hw_rule_add(runtime, cond, var: -1, |
2479 | func: snd_pcm_hw_rule_subformats, private: (void *)subformats, |
2480 | SNDRV_PCM_HW_PARAM_SUBFORMAT, |
2481 | SNDRV_PCM_HW_PARAM_FORMAT, -1); |
2482 | } |
2483 | |
2484 | static int snd_pcm_hw_constraints_init(struct snd_pcm_substream *substream) |
2485 | { |
2486 | struct snd_pcm_runtime *runtime = substream->runtime; |
2487 | struct snd_pcm_hw_constraints *constrs = &runtime->hw_constraints; |
2488 | int k, err; |
2489 | |
2490 | for (k = SNDRV_PCM_HW_PARAM_FIRST_MASK; k <= SNDRV_PCM_HW_PARAM_LAST_MASK; k++) { |
2491 | snd_mask_any(mask: constrs_mask(constrs, var: k)); |
2492 | } |
2493 | |
2494 | for (k = SNDRV_PCM_HW_PARAM_FIRST_INTERVAL; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++) { |
2495 | snd_interval_any(i: constrs_interval(constrs, var: k)); |
2496 | } |
2497 | |
2498 | snd_interval_setinteger(i: constrs_interval(constrs, SNDRV_PCM_HW_PARAM_CHANNELS)); |
2499 | snd_interval_setinteger(i: constrs_interval(constrs, SNDRV_PCM_HW_PARAM_BUFFER_SIZE)); |
2500 | snd_interval_setinteger(i: constrs_interval(constrs, SNDRV_PCM_HW_PARAM_BUFFER_BYTES)); |
2501 | snd_interval_setinteger(i: constrs_interval(constrs, SNDRV_PCM_HW_PARAM_SAMPLE_BITS)); |
2502 | snd_interval_setinteger(i: constrs_interval(constrs, SNDRV_PCM_HW_PARAM_FRAME_BITS)); |
2503 | |
2504 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_FORMAT, |
2505 | func: snd_pcm_hw_rule_format, NULL, |
2506 | SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1); |
2507 | if (err < 0) |
2508 | return err; |
2509 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_SAMPLE_BITS, |
2510 | func: snd_pcm_hw_rule_sample_bits, NULL, |
2511 | SNDRV_PCM_HW_PARAM_FORMAT, |
2512 | SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1); |
2513 | if (err < 0) |
2514 | return err; |
2515 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_SAMPLE_BITS, |
2516 | func: snd_pcm_hw_rule_div, NULL, |
2517 | SNDRV_PCM_HW_PARAM_FRAME_BITS, SNDRV_PCM_HW_PARAM_CHANNELS, -1); |
2518 | if (err < 0) |
2519 | return err; |
2520 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_FRAME_BITS, |
2521 | func: snd_pcm_hw_rule_mul, NULL, |
2522 | SNDRV_PCM_HW_PARAM_SAMPLE_BITS, SNDRV_PCM_HW_PARAM_CHANNELS, -1); |
2523 | if (err < 0) |
2524 | return err; |
2525 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_FRAME_BITS, |
2526 | func: snd_pcm_hw_rule_mulkdiv, private: (void*) 8, |
2527 | SNDRV_PCM_HW_PARAM_PERIOD_BYTES, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, -1); |
2528 | if (err < 0) |
2529 | return err; |
2530 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_FRAME_BITS, |
2531 | func: snd_pcm_hw_rule_mulkdiv, private: (void*) 8, |
2532 | SNDRV_PCM_HW_PARAM_BUFFER_BYTES, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, -1); |
2533 | if (err < 0) |
2534 | return err; |
2535 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_CHANNELS, |
2536 | func: snd_pcm_hw_rule_div, NULL, |
2537 | SNDRV_PCM_HW_PARAM_FRAME_BITS, SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1); |
2538 | if (err < 0) |
2539 | return err; |
2540 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_RATE, |
2541 | func: snd_pcm_hw_rule_mulkdiv, private: (void*) 1000000, |
2542 | SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_PERIOD_TIME, -1); |
2543 | if (err < 0) |
2544 | return err; |
2545 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_RATE, |
2546 | func: snd_pcm_hw_rule_mulkdiv, private: (void*) 1000000, |
2547 | SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_BUFFER_TIME, -1); |
2548 | if (err < 0) |
2549 | return err; |
2550 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_PERIODS, |
2551 | func: snd_pcm_hw_rule_div, NULL, |
2552 | SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, -1); |
2553 | if (err < 0) |
2554 | return err; |
2555 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, |
2556 | func: snd_pcm_hw_rule_div, NULL, |
2557 | SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_PERIODS, -1); |
2558 | if (err < 0) |
2559 | return err; |
2560 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, |
2561 | func: snd_pcm_hw_rule_mulkdiv, private: (void*) 8, |
2562 | SNDRV_PCM_HW_PARAM_PERIOD_BYTES, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1); |
2563 | if (err < 0) |
2564 | return err; |
2565 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, |
2566 | func: snd_pcm_hw_rule_muldivk, private: (void*) 1000000, |
2567 | SNDRV_PCM_HW_PARAM_PERIOD_TIME, SNDRV_PCM_HW_PARAM_RATE, -1); |
2568 | if (err < 0) |
2569 | return err; |
2570 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, |
2571 | func: snd_pcm_hw_rule_mul, NULL, |
2572 | SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_PERIODS, -1); |
2573 | if (err < 0) |
2574 | return err; |
2575 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, |
2576 | func: snd_pcm_hw_rule_mulkdiv, private: (void*) 8, |
2577 | SNDRV_PCM_HW_PARAM_BUFFER_BYTES, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1); |
2578 | if (err < 0) |
2579 | return err; |
2580 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, |
2581 | func: snd_pcm_hw_rule_muldivk, private: (void*) 1000000, |
2582 | SNDRV_PCM_HW_PARAM_BUFFER_TIME, SNDRV_PCM_HW_PARAM_RATE, -1); |
2583 | if (err < 0) |
2584 | return err; |
2585 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_PERIOD_BYTES, |
2586 | func: snd_pcm_hw_rule_muldivk, private: (void*) 8, |
2587 | SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1); |
2588 | if (err < 0) |
2589 | return err; |
2590 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, |
2591 | func: snd_pcm_hw_rule_muldivk, private: (void*) 8, |
2592 | SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1); |
2593 | if (err < 0) |
2594 | return err; |
2595 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_PERIOD_TIME, |
2596 | func: snd_pcm_hw_rule_mulkdiv, private: (void*) 1000000, |
2597 | SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_RATE, -1); |
2598 | if (err < 0) |
2599 | return err; |
2600 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_BUFFER_TIME, |
2601 | func: snd_pcm_hw_rule_mulkdiv, private: (void*) 1000000, |
2602 | SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_RATE, -1); |
2603 | if (err < 0) |
2604 | return err; |
2605 | return 0; |
2606 | } |
2607 | |
2608 | static int snd_pcm_hw_constraints_complete(struct snd_pcm_substream *substream) |
2609 | { |
2610 | struct snd_pcm_runtime *runtime = substream->runtime; |
2611 | struct snd_pcm_hardware *hw = &runtime->hw; |
2612 | int err; |
2613 | unsigned int mask = 0; |
2614 | |
2615 | if (hw->info & SNDRV_PCM_INFO_INTERLEAVED) |
2616 | mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_RW_INTERLEAVED); |
2617 | if (hw->info & SNDRV_PCM_INFO_NONINTERLEAVED) |
2618 | mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_RW_NONINTERLEAVED); |
2619 | if (hw_support_mmap(substream)) { |
2620 | if (hw->info & SNDRV_PCM_INFO_INTERLEAVED) |
2621 | mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_MMAP_INTERLEAVED); |
2622 | if (hw->info & SNDRV_PCM_INFO_NONINTERLEAVED) |
2623 | mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_MMAP_NONINTERLEAVED); |
2624 | if (hw->info & SNDRV_PCM_INFO_COMPLEX) |
2625 | mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_MMAP_COMPLEX); |
2626 | } |
2627 | err = snd_pcm_hw_constraint_mask(runtime, SNDRV_PCM_HW_PARAM_ACCESS, mask); |
2628 | if (err < 0) |
2629 | return err; |
2630 | |
2631 | err = snd_pcm_hw_constraint_mask64(runtime, SNDRV_PCM_HW_PARAM_FORMAT, mask: hw->formats); |
2632 | if (err < 0) |
2633 | return err; |
2634 | |
2635 | err = snd_pcm_hw_constraint_subformats(runtime, cond: 0, subformats: &hw->subformats); |
2636 | if (err < 0) |
2637 | return err; |
2638 | |
2639 | err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_CHANNELS, |
2640 | min: hw->channels_min, max: hw->channels_max); |
2641 | if (err < 0) |
2642 | return err; |
2643 | |
2644 | err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_RATE, |
2645 | min: hw->rate_min, max: hw->rate_max); |
2646 | if (err < 0) |
2647 | return err; |
2648 | |
2649 | err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_PERIOD_BYTES, |
2650 | min: hw->period_bytes_min, max: hw->period_bytes_max); |
2651 | if (err < 0) |
2652 | return err; |
2653 | |
2654 | err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_PERIODS, |
2655 | min: hw->periods_min, max: hw->periods_max); |
2656 | if (err < 0) |
2657 | return err; |
2658 | |
2659 | err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, |
2660 | min: hw->period_bytes_min, max: hw->buffer_bytes_max); |
2661 | if (err < 0) |
2662 | return err; |
2663 | |
2664 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, |
2665 | func: snd_pcm_hw_rule_buffer_bytes_max, private: substream, |
2666 | SNDRV_PCM_HW_PARAM_BUFFER_BYTES, -1); |
2667 | if (err < 0) |
2668 | return err; |
2669 | |
2670 | /* FIXME: remove */ |
2671 | if (runtime->dma_bytes) { |
2672 | err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, min: 0, max: runtime->dma_bytes); |
2673 | if (err < 0) |
2674 | return err; |
2675 | } |
2676 | |
2677 | if (!(hw->rates & (SNDRV_PCM_RATE_KNOT | SNDRV_PCM_RATE_CONTINUOUS))) { |
2678 | err = snd_pcm_hw_rule_add(runtime, cond: 0, SNDRV_PCM_HW_PARAM_RATE, |
2679 | func: snd_pcm_hw_rule_rate, private: hw, |
2680 | SNDRV_PCM_HW_PARAM_RATE, -1); |
2681 | if (err < 0) |
2682 | return err; |
2683 | } |
2684 | |
2685 | /* FIXME: this belong to lowlevel */ |
2686 | snd_pcm_hw_constraint_integer(runtime, SNDRV_PCM_HW_PARAM_PERIOD_SIZE); |
2687 | |
2688 | return 0; |
2689 | } |
2690 | |
2691 | static void pcm_release_private(struct snd_pcm_substream *substream) |
2692 | { |
2693 | if (snd_pcm_stream_linked(substream)) |
2694 | snd_pcm_unlink(substream); |
2695 | } |
2696 | |
2697 | void snd_pcm_release_substream(struct snd_pcm_substream *substream) |
2698 | { |
2699 | substream->ref_count--; |
2700 | if (substream->ref_count > 0) |
2701 | return; |
2702 | |
2703 | snd_pcm_drop(substream); |
2704 | if (substream->hw_opened) { |
2705 | if (substream->runtime->state != SNDRV_PCM_STATE_OPEN) |
2706 | do_hw_free(substream); |
2707 | substream->ops->close(substream); |
2708 | substream->hw_opened = 0; |
2709 | } |
2710 | if (cpu_latency_qos_request_active(req: &substream->latency_pm_qos_req)) |
2711 | cpu_latency_qos_remove_request(req: &substream->latency_pm_qos_req); |
2712 | if (substream->pcm_release) { |
2713 | substream->pcm_release(substream); |
2714 | substream->pcm_release = NULL; |
2715 | } |
2716 | snd_pcm_detach_substream(substream); |
2717 | } |
2718 | EXPORT_SYMBOL(snd_pcm_release_substream); |
2719 | |
2720 | int snd_pcm_open_substream(struct snd_pcm *pcm, int stream, |
2721 | struct file *file, |
2722 | struct snd_pcm_substream **rsubstream) |
2723 | { |
2724 | struct snd_pcm_substream *substream; |
2725 | int err; |
2726 | |
2727 | err = snd_pcm_attach_substream(pcm, stream, file, rsubstream: &substream); |
2728 | if (err < 0) |
2729 | return err; |
2730 | if (substream->ref_count > 1) { |
2731 | *rsubstream = substream; |
2732 | return 0; |
2733 | } |
2734 | |
2735 | err = snd_pcm_hw_constraints_init(substream); |
2736 | if (err < 0) { |
2737 | pcm_dbg(pcm, "snd_pcm_hw_constraints_init failed\n" ); |
2738 | goto error; |
2739 | } |
2740 | |
2741 | err = substream->ops->open(substream); |
2742 | if (err < 0) |
2743 | goto error; |
2744 | |
2745 | substream->hw_opened = 1; |
2746 | |
2747 | err = snd_pcm_hw_constraints_complete(substream); |
2748 | if (err < 0) { |
2749 | pcm_dbg(pcm, "snd_pcm_hw_constraints_complete failed\n" ); |
2750 | goto error; |
2751 | } |
2752 | |
2753 | /* automatically set EXPLICIT_SYNC flag in the managed mode whenever |
2754 | * the DMA buffer requires it |
2755 | */ |
2756 | if (substream->managed_buffer_alloc && |
2757 | substream->dma_buffer.dev.need_sync) |
2758 | substream->runtime->hw.info |= SNDRV_PCM_INFO_EXPLICIT_SYNC; |
2759 | |
2760 | *rsubstream = substream; |
2761 | return 0; |
2762 | |
2763 | error: |
2764 | snd_pcm_release_substream(substream); |
2765 | return err; |
2766 | } |
2767 | EXPORT_SYMBOL(snd_pcm_open_substream); |
2768 | |
2769 | static int snd_pcm_open_file(struct file *file, |
2770 | struct snd_pcm *pcm, |
2771 | int stream) |
2772 | { |
2773 | struct snd_pcm_file *pcm_file; |
2774 | struct snd_pcm_substream *substream; |
2775 | int err; |
2776 | |
2777 | err = snd_pcm_open_substream(pcm, stream, file, &substream); |
2778 | if (err < 0) |
2779 | return err; |
2780 | |
2781 | pcm_file = kzalloc(size: sizeof(*pcm_file), GFP_KERNEL); |
2782 | if (pcm_file == NULL) { |
2783 | snd_pcm_release_substream(substream); |
2784 | return -ENOMEM; |
2785 | } |
2786 | pcm_file->substream = substream; |
2787 | if (substream->ref_count == 1) |
2788 | substream->pcm_release = pcm_release_private; |
2789 | file->private_data = pcm_file; |
2790 | |
2791 | return 0; |
2792 | } |
2793 | |
2794 | static int snd_pcm_playback_open(struct inode *inode, struct file *file) |
2795 | { |
2796 | struct snd_pcm *pcm; |
2797 | int err = nonseekable_open(inode, filp: file); |
2798 | if (err < 0) |
2799 | return err; |
2800 | pcm = snd_lookup_minor_data(minor: iminor(inode), |
2801 | type: SNDRV_DEVICE_TYPE_PCM_PLAYBACK); |
2802 | err = snd_pcm_open(file, pcm, stream: SNDRV_PCM_STREAM_PLAYBACK); |
2803 | if (pcm) |
2804 | snd_card_unref(card: pcm->card); |
2805 | return err; |
2806 | } |
2807 | |
2808 | static int snd_pcm_capture_open(struct inode *inode, struct file *file) |
2809 | { |
2810 | struct snd_pcm *pcm; |
2811 | int err = nonseekable_open(inode, filp: file); |
2812 | if (err < 0) |
2813 | return err; |
2814 | pcm = snd_lookup_minor_data(minor: iminor(inode), |
2815 | type: SNDRV_DEVICE_TYPE_PCM_CAPTURE); |
2816 | err = snd_pcm_open(file, pcm, stream: SNDRV_PCM_STREAM_CAPTURE); |
2817 | if (pcm) |
2818 | snd_card_unref(card: pcm->card); |
2819 | return err; |
2820 | } |
2821 | |
2822 | static int snd_pcm_open(struct file *file, struct snd_pcm *pcm, int stream) |
2823 | { |
2824 | int err; |
2825 | wait_queue_entry_t wait; |
2826 | |
2827 | if (pcm == NULL) { |
2828 | err = -ENODEV; |
2829 | goto __error1; |
2830 | } |
2831 | err = snd_card_file_add(card: pcm->card, file); |
2832 | if (err < 0) |
2833 | goto __error1; |
2834 | if (!try_module_get(module: pcm->card->module)) { |
2835 | err = -EFAULT; |
2836 | goto __error2; |
2837 | } |
2838 | init_waitqueue_entry(wq_entry: &wait, current); |
2839 | add_wait_queue(wq_head: &pcm->open_wait, wq_entry: &wait); |
2840 | mutex_lock(&pcm->open_mutex); |
2841 | while (1) { |
2842 | err = snd_pcm_open_file(file, pcm, stream); |
2843 | if (err >= 0) |
2844 | break; |
2845 | if (err == -EAGAIN) { |
2846 | if (file->f_flags & O_NONBLOCK) { |
2847 | err = -EBUSY; |
2848 | break; |
2849 | } |
2850 | } else |
2851 | break; |
2852 | set_current_state(TASK_INTERRUPTIBLE); |
2853 | mutex_unlock(lock: &pcm->open_mutex); |
2854 | schedule(); |
2855 | mutex_lock(&pcm->open_mutex); |
2856 | if (pcm->card->shutdown) { |
2857 | err = -ENODEV; |
2858 | break; |
2859 | } |
2860 | if (signal_pending(current)) { |
2861 | err = -ERESTARTSYS; |
2862 | break; |
2863 | } |
2864 | } |
2865 | remove_wait_queue(wq_head: &pcm->open_wait, wq_entry: &wait); |
2866 | mutex_unlock(lock: &pcm->open_mutex); |
2867 | if (err < 0) |
2868 | goto __error; |
2869 | return err; |
2870 | |
2871 | __error: |
2872 | module_put(module: pcm->card->module); |
2873 | __error2: |
2874 | snd_card_file_remove(card: pcm->card, file); |
2875 | __error1: |
2876 | return err; |
2877 | } |
2878 | |
2879 | static int snd_pcm_release(struct inode *inode, struct file *file) |
2880 | { |
2881 | struct snd_pcm *pcm; |
2882 | struct snd_pcm_substream *substream; |
2883 | struct snd_pcm_file *pcm_file; |
2884 | |
2885 | pcm_file = file->private_data; |
2886 | substream = pcm_file->substream; |
2887 | if (snd_BUG_ON(!substream)) |
2888 | return -ENXIO; |
2889 | pcm = substream->pcm; |
2890 | |
2891 | /* block until the device gets woken up as it may touch the hardware */ |
2892 | snd_power_wait(card: pcm->card); |
2893 | |
2894 | scoped_guard(mutex, &pcm->open_mutex) { |
2895 | snd_pcm_release_substream(substream); |
2896 | kfree(objp: pcm_file); |
2897 | } |
2898 | wake_up(&pcm->open_wait); |
2899 | module_put(module: pcm->card->module); |
2900 | snd_card_file_remove(card: pcm->card, file); |
2901 | return 0; |
2902 | } |
2903 | |
2904 | /* check and update PCM state; return 0 or a negative error |
2905 | * call this inside PCM lock |
2906 | */ |
2907 | static int do_pcm_hwsync(struct snd_pcm_substream *substream) |
2908 | { |
2909 | switch (substream->runtime->state) { |
2910 | case SNDRV_PCM_STATE_DRAINING: |
2911 | if (substream->stream == SNDRV_PCM_STREAM_CAPTURE) |
2912 | return -EBADFD; |
2913 | fallthrough; |
2914 | case SNDRV_PCM_STATE_RUNNING: |
2915 | return snd_pcm_update_hw_ptr(substream); |
2916 | case SNDRV_PCM_STATE_PREPARED: |
2917 | case SNDRV_PCM_STATE_PAUSED: |
2918 | return 0; |
2919 | case SNDRV_PCM_STATE_SUSPENDED: |
2920 | return -ESTRPIPE; |
2921 | case SNDRV_PCM_STATE_XRUN: |
2922 | return -EPIPE; |
2923 | default: |
2924 | return -EBADFD; |
2925 | } |
2926 | } |
2927 | |
2928 | /* increase the appl_ptr; returns the processed frames or a negative error */ |
2929 | static snd_pcm_sframes_t forward_appl_ptr(struct snd_pcm_substream *substream, |
2930 | snd_pcm_uframes_t frames, |
2931 | snd_pcm_sframes_t avail) |
2932 | { |
2933 | struct snd_pcm_runtime *runtime = substream->runtime; |
2934 | snd_pcm_sframes_t appl_ptr; |
2935 | int ret; |
2936 | |
2937 | if (avail <= 0) |
2938 | return 0; |
2939 | if (frames > (snd_pcm_uframes_t)avail) |
2940 | frames = avail; |
2941 | appl_ptr = runtime->control->appl_ptr + frames; |
2942 | if (appl_ptr >= (snd_pcm_sframes_t)runtime->boundary) |
2943 | appl_ptr -= runtime->boundary; |
2944 | ret = pcm_lib_apply_appl_ptr(substream, appl_ptr); |
2945 | return ret < 0 ? ret : frames; |
2946 | } |
2947 | |
2948 | /* decrease the appl_ptr; returns the processed frames or zero for error */ |
2949 | static snd_pcm_sframes_t rewind_appl_ptr(struct snd_pcm_substream *substream, |
2950 | snd_pcm_uframes_t frames, |
2951 | snd_pcm_sframes_t avail) |
2952 | { |
2953 | struct snd_pcm_runtime *runtime = substream->runtime; |
2954 | snd_pcm_sframes_t appl_ptr; |
2955 | int ret; |
2956 | |
2957 | if (avail <= 0) |
2958 | return 0; |
2959 | if (frames > (snd_pcm_uframes_t)avail) |
2960 | frames = avail; |
2961 | appl_ptr = runtime->control->appl_ptr - frames; |
2962 | if (appl_ptr < 0) |
2963 | appl_ptr += runtime->boundary; |
2964 | ret = pcm_lib_apply_appl_ptr(substream, appl_ptr); |
2965 | /* NOTE: we return zero for errors because PulseAudio gets depressed |
2966 | * upon receiving an error from rewind ioctl and stops processing |
2967 | * any longer. Returning zero means that no rewind is done, so |
2968 | * it's not absolutely wrong to answer like that. |
2969 | */ |
2970 | return ret < 0 ? 0 : frames; |
2971 | } |
2972 | |
2973 | static snd_pcm_sframes_t snd_pcm_rewind(struct snd_pcm_substream *substream, |
2974 | snd_pcm_uframes_t frames) |
2975 | { |
2976 | snd_pcm_sframes_t ret; |
2977 | |
2978 | if (frames == 0) |
2979 | return 0; |
2980 | |
2981 | scoped_guard(pcm_stream_lock_irq, substream) { |
2982 | ret = do_pcm_hwsync(substream); |
2983 | if (!ret) |
2984 | ret = rewind_appl_ptr(substream, frames, |
2985 | avail: snd_pcm_hw_avail(substream)); |
2986 | } |
2987 | if (ret >= 0) |
2988 | snd_pcm_dma_buffer_sync(substream, mode: SNDRV_DMA_SYNC_DEVICE); |
2989 | return ret; |
2990 | } |
2991 | |
2992 | static snd_pcm_sframes_t snd_pcm_forward(struct snd_pcm_substream *substream, |
2993 | snd_pcm_uframes_t frames) |
2994 | { |
2995 | snd_pcm_sframes_t ret; |
2996 | |
2997 | if (frames == 0) |
2998 | return 0; |
2999 | |
3000 | scoped_guard(pcm_stream_lock_irq, substream) { |
3001 | ret = do_pcm_hwsync(substream); |
3002 | if (!ret) |
3003 | ret = forward_appl_ptr(substream, frames, |
3004 | avail: snd_pcm_avail(substream)); |
3005 | } |
3006 | if (ret >= 0) |
3007 | snd_pcm_dma_buffer_sync(substream, mode: SNDRV_DMA_SYNC_DEVICE); |
3008 | return ret; |
3009 | } |
3010 | |
3011 | static int snd_pcm_delay(struct snd_pcm_substream *substream, |
3012 | snd_pcm_sframes_t *delay) |
3013 | { |
3014 | int err; |
3015 | |
3016 | scoped_guard(pcm_stream_lock_irq, substream) { |
3017 | err = do_pcm_hwsync(substream); |
3018 | if (delay && !err) |
3019 | *delay = snd_pcm_calc_delay(substream); |
3020 | } |
3021 | snd_pcm_dma_buffer_sync(substream, mode: SNDRV_DMA_SYNC_CPU); |
3022 | |
3023 | return err; |
3024 | } |
3025 | |
3026 | static inline int snd_pcm_hwsync(struct snd_pcm_substream *substream) |
3027 | { |
3028 | return snd_pcm_delay(substream, NULL); |
3029 | } |
3030 | |
3031 | static int snd_pcm_sync_ptr(struct snd_pcm_substream *substream, |
3032 | struct snd_pcm_sync_ptr __user *_sync_ptr) |
3033 | { |
3034 | struct snd_pcm_runtime *runtime = substream->runtime; |
3035 | struct snd_pcm_sync_ptr sync_ptr; |
3036 | volatile struct snd_pcm_mmap_status *status; |
3037 | volatile struct snd_pcm_mmap_control *control; |
3038 | int err; |
3039 | |
3040 | memset(&sync_ptr, 0, sizeof(sync_ptr)); |
3041 | if (get_user(sync_ptr.flags, (unsigned __user *)&(_sync_ptr->flags))) |
3042 | return -EFAULT; |
3043 | if (copy_from_user(to: &sync_ptr.c.control, from: &(_sync_ptr->c.control), n: sizeof(struct snd_pcm_mmap_control))) |
3044 | return -EFAULT; |
3045 | status = runtime->status; |
3046 | control = runtime->control; |
3047 | if (sync_ptr.flags & SNDRV_PCM_SYNC_PTR_HWSYNC) { |
3048 | err = snd_pcm_hwsync(substream); |
3049 | if (err < 0) |
3050 | return err; |
3051 | } |
3052 | scoped_guard(pcm_stream_lock_irq, substream) { |
3053 | if (!(sync_ptr.flags & SNDRV_PCM_SYNC_PTR_APPL)) { |
3054 | err = pcm_lib_apply_appl_ptr(substream, |
3055 | appl_ptr: sync_ptr.c.control.appl_ptr); |
3056 | if (err < 0) |
3057 | return err; |
3058 | } else { |
3059 | sync_ptr.c.control.appl_ptr = control->appl_ptr; |
3060 | } |
3061 | if (!(sync_ptr.flags & SNDRV_PCM_SYNC_PTR_AVAIL_MIN)) |
3062 | control->avail_min = sync_ptr.c.control.avail_min; |
3063 | else |
3064 | sync_ptr.c.control.avail_min = control->avail_min; |
3065 | sync_ptr.s.status.state = status->state; |
3066 | sync_ptr.s.status.hw_ptr = status->hw_ptr; |
3067 | sync_ptr.s.status.tstamp = status->tstamp; |
3068 | sync_ptr.s.status.suspended_state = status->suspended_state; |
3069 | sync_ptr.s.status.audio_tstamp = status->audio_tstamp; |
3070 | } |
3071 | if (!(sync_ptr.flags & SNDRV_PCM_SYNC_PTR_APPL)) |
3072 | snd_pcm_dma_buffer_sync(substream, mode: SNDRV_DMA_SYNC_DEVICE); |
3073 | if (copy_to_user(to: _sync_ptr, from: &sync_ptr, n: sizeof(sync_ptr))) |
3074 | return -EFAULT; |
3075 | return 0; |
3076 | } |
3077 | |
3078 | struct snd_pcm_mmap_status32 { |
3079 | snd_pcm_state_t state; |
3080 | s32 pad1; |
3081 | u32 hw_ptr; |
3082 | s32 tstamp_sec; |
3083 | s32 tstamp_nsec; |
3084 | snd_pcm_state_t suspended_state; |
3085 | s32 audio_tstamp_sec; |
3086 | s32 audio_tstamp_nsec; |
3087 | } __packed; |
3088 | |
3089 | struct snd_pcm_mmap_control32 { |
3090 | u32 appl_ptr; |
3091 | u32 avail_min; |
3092 | }; |
3093 | |
3094 | struct snd_pcm_sync_ptr32 { |
3095 | u32 flags; |
3096 | union { |
3097 | struct snd_pcm_mmap_status32 status; |
3098 | unsigned char reserved[64]; |
3099 | } s; |
3100 | union { |
3101 | struct snd_pcm_mmap_control32 control; |
3102 | unsigned char reserved[64]; |
3103 | } c; |
3104 | } __packed; |
3105 | |
3106 | /* recalcuate the boundary within 32bit */ |
3107 | static snd_pcm_uframes_t recalculate_boundary(struct snd_pcm_runtime *runtime) |
3108 | { |
3109 | snd_pcm_uframes_t boundary; |
3110 | |
3111 | if (! runtime->buffer_size) |
3112 | return 0; |
3113 | boundary = runtime->buffer_size; |
3114 | while (boundary * 2 <= 0x7fffffffUL - runtime->buffer_size) |
3115 | boundary *= 2; |
3116 | return boundary; |
3117 | } |
3118 | |
3119 | static int snd_pcm_ioctl_sync_ptr_compat(struct snd_pcm_substream *substream, |
3120 | struct snd_pcm_sync_ptr32 __user *src) |
3121 | { |
3122 | struct snd_pcm_runtime *runtime = substream->runtime; |
3123 | volatile struct snd_pcm_mmap_status *status; |
3124 | volatile struct snd_pcm_mmap_control *control; |
3125 | u32 sflags; |
3126 | struct snd_pcm_mmap_control scontrol; |
3127 | struct snd_pcm_mmap_status sstatus; |
3128 | snd_pcm_uframes_t boundary; |
3129 | int err; |
3130 | |
3131 | if (snd_BUG_ON(!runtime)) |
3132 | return -EINVAL; |
3133 | |
3134 | if (get_user(sflags, &src->flags) || |
3135 | get_user(scontrol.appl_ptr, &src->c.control.appl_ptr) || |
3136 | get_user(scontrol.avail_min, &src->c.control.avail_min)) |
3137 | return -EFAULT; |
3138 | if (sflags & SNDRV_PCM_SYNC_PTR_HWSYNC) { |
3139 | err = snd_pcm_hwsync(substream); |
3140 | if (err < 0) |
3141 | return err; |
3142 | } |
3143 | status = runtime->status; |
3144 | control = runtime->control; |
3145 | boundary = recalculate_boundary(runtime); |
3146 | if (! boundary) |
3147 | boundary = 0x7fffffff; |
3148 | scoped_guard(pcm_stream_lock_irq, substream) { |
3149 | /* FIXME: we should consider the boundary for the sync from app */ |
3150 | if (!(sflags & SNDRV_PCM_SYNC_PTR_APPL)) { |
3151 | err = pcm_lib_apply_appl_ptr(substream, |
3152 | appl_ptr: scontrol.appl_ptr); |
3153 | if (err < 0) |
3154 | return err; |
3155 | } else |
3156 | scontrol.appl_ptr = control->appl_ptr % boundary; |
3157 | if (!(sflags & SNDRV_PCM_SYNC_PTR_AVAIL_MIN)) |
3158 | control->avail_min = scontrol.avail_min; |
3159 | else |
3160 | scontrol.avail_min = control->avail_min; |
3161 | sstatus.state = status->state; |
3162 | sstatus.hw_ptr = status->hw_ptr % boundary; |
3163 | sstatus.tstamp = status->tstamp; |
3164 | sstatus.suspended_state = status->suspended_state; |
3165 | sstatus.audio_tstamp = status->audio_tstamp; |
3166 | } |
3167 | if (!(sflags & SNDRV_PCM_SYNC_PTR_APPL)) |
3168 | snd_pcm_dma_buffer_sync(substream, mode: SNDRV_DMA_SYNC_DEVICE); |
3169 | if (put_user(sstatus.state, &src->s.status.state) || |
3170 | put_user(sstatus.hw_ptr, &src->s.status.hw_ptr) || |
3171 | put_user(sstatus.tstamp.tv_sec, &src->s.status.tstamp_sec) || |
3172 | put_user(sstatus.tstamp.tv_nsec, &src->s.status.tstamp_nsec) || |
3173 | put_user(sstatus.suspended_state, &src->s.status.suspended_state) || |
3174 | put_user(sstatus.audio_tstamp.tv_sec, &src->s.status.audio_tstamp_sec) || |
3175 | put_user(sstatus.audio_tstamp.tv_nsec, &src->s.status.audio_tstamp_nsec) || |
3176 | put_user(scontrol.appl_ptr, &src->c.control.appl_ptr) || |
3177 | put_user(scontrol.avail_min, &src->c.control.avail_min)) |
3178 | return -EFAULT; |
3179 | |
3180 | return 0; |
3181 | } |
3182 | #define __SNDRV_PCM_IOCTL_SYNC_PTR32 _IOWR('A', 0x23, struct snd_pcm_sync_ptr32) |
3183 | |
3184 | static int snd_pcm_tstamp(struct snd_pcm_substream *substream, int __user *_arg) |
3185 | { |
3186 | struct snd_pcm_runtime *runtime = substream->runtime; |
3187 | int arg; |
3188 | |
3189 | if (get_user(arg, _arg)) |
3190 | return -EFAULT; |
3191 | if (arg < 0 || arg > SNDRV_PCM_TSTAMP_TYPE_LAST) |
3192 | return -EINVAL; |
3193 | runtime->tstamp_type = arg; |
3194 | return 0; |
3195 | } |
3196 | |
3197 | static int snd_pcm_xferi_frames_ioctl(struct snd_pcm_substream *substream, |
3198 | struct snd_xferi __user *_xferi) |
3199 | { |
3200 | struct snd_xferi xferi; |
3201 | struct snd_pcm_runtime *runtime = substream->runtime; |
3202 | snd_pcm_sframes_t result; |
3203 | |
3204 | if (runtime->state == SNDRV_PCM_STATE_OPEN) |
3205 | return -EBADFD; |
3206 | if (put_user(0, &_xferi->result)) |
3207 | return -EFAULT; |
3208 | if (copy_from_user(to: &xferi, from: _xferi, n: sizeof(xferi))) |
3209 | return -EFAULT; |
3210 | if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) |
3211 | result = snd_pcm_lib_write(substream, buf: xferi.buf, frames: xferi.frames); |
3212 | else |
3213 | result = snd_pcm_lib_read(substream, buf: xferi.buf, frames: xferi.frames); |
3214 | if (put_user(result, &_xferi->result)) |
3215 | return -EFAULT; |
3216 | return result < 0 ? result : 0; |
3217 | } |
3218 | |
3219 | static int snd_pcm_xfern_frames_ioctl(struct snd_pcm_substream *substream, |
3220 | struct snd_xfern __user *_xfern) |
3221 | { |
3222 | struct snd_xfern xfern; |
3223 | struct snd_pcm_runtime *runtime = substream->runtime; |
3224 | void *bufs __free(kfree) = NULL; |
3225 | snd_pcm_sframes_t result; |
3226 | |
3227 | if (runtime->state == SNDRV_PCM_STATE_OPEN) |
3228 | return -EBADFD; |
3229 | if (runtime->channels > 128) |
3230 | return -EINVAL; |
3231 | if (put_user(0, &_xfern->result)) |
3232 | return -EFAULT; |
3233 | if (copy_from_user(to: &xfern, from: _xfern, n: sizeof(xfern))) |
3234 | return -EFAULT; |
3235 | |
3236 | bufs = memdup_user(xfern.bufs, sizeof(void *) * runtime->channels); |
3237 | if (IS_ERR(ptr: bufs)) |
3238 | return PTR_ERR(no_free_ptr(bufs)); |
3239 | if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) |
3240 | result = snd_pcm_lib_writev(substream, bufs, frames: xfern.frames); |
3241 | else |
3242 | result = snd_pcm_lib_readv(substream, bufs, frames: xfern.frames); |
3243 | if (put_user(result, &_xfern->result)) |
3244 | return -EFAULT; |
3245 | return result < 0 ? result : 0; |
3246 | } |
3247 | |
3248 | static int snd_pcm_rewind_ioctl(struct snd_pcm_substream *substream, |
3249 | snd_pcm_uframes_t __user *_frames) |
3250 | { |
3251 | snd_pcm_uframes_t frames; |
3252 | snd_pcm_sframes_t result; |
3253 | |
3254 | if (get_user(frames, _frames)) |
3255 | return -EFAULT; |
3256 | if (put_user(0, _frames)) |
3257 | return -EFAULT; |
3258 | result = snd_pcm_rewind(substream, frames); |
3259 | if (put_user(result, _frames)) |
3260 | return -EFAULT; |
3261 | return result < 0 ? result : 0; |
3262 | } |
3263 | |
3264 | static int snd_pcm_forward_ioctl(struct snd_pcm_substream *substream, |
3265 | snd_pcm_uframes_t __user *_frames) |
3266 | { |
3267 | snd_pcm_uframes_t frames; |
3268 | snd_pcm_sframes_t result; |
3269 | |
3270 | if (get_user(frames, _frames)) |
3271 | return -EFAULT; |
3272 | if (put_user(0, _frames)) |
3273 | return -EFAULT; |
3274 | result = snd_pcm_forward(substream, frames); |
3275 | if (put_user(result, _frames)) |
3276 | return -EFAULT; |
3277 | return result < 0 ? result : 0; |
3278 | } |
3279 | |
3280 | static int snd_pcm_common_ioctl(struct file *file, |
3281 | struct snd_pcm_substream *substream, |
3282 | unsigned int cmd, void __user *arg) |
3283 | { |
3284 | struct snd_pcm_file *pcm_file = file->private_data; |
3285 | int res; |
3286 | |
3287 | if (PCM_RUNTIME_CHECK(substream)) |
3288 | return -ENXIO; |
3289 | |
3290 | if (substream->runtime->state == SNDRV_PCM_STATE_DISCONNECTED) |
3291 | return -EBADFD; |
3292 | |
3293 | res = snd_power_wait(card: substream->pcm->card); |
3294 | if (res < 0) |
3295 | return res; |
3296 | |
3297 | switch (cmd) { |
3298 | case SNDRV_PCM_IOCTL_PVERSION: |
3299 | return put_user(SNDRV_PCM_VERSION, (int __user *)arg) ? -EFAULT : 0; |
3300 | case SNDRV_PCM_IOCTL_INFO: |
3301 | return snd_pcm_info_user(substream, info: arg); |
3302 | case SNDRV_PCM_IOCTL_TSTAMP: /* just for compatibility */ |
3303 | return 0; |
3304 | case SNDRV_PCM_IOCTL_TTSTAMP: |
3305 | return snd_pcm_tstamp(substream, arg: arg); |
3306 | case SNDRV_PCM_IOCTL_USER_PVERSION: |
3307 | if (get_user(pcm_file->user_pversion, |
3308 | (unsigned int __user *)arg)) |
3309 | return -EFAULT; |
3310 | return 0; |
3311 | case SNDRV_PCM_IOCTL_HW_REFINE: |
3312 | return snd_pcm_hw_refine_user(substream, params: arg); |
3313 | case SNDRV_PCM_IOCTL_HW_PARAMS: |
3314 | return snd_pcm_hw_params_user(substream, params: arg); |
3315 | case SNDRV_PCM_IOCTL_HW_FREE: |
3316 | return snd_pcm_hw_free(substream); |
3317 | case SNDRV_PCM_IOCTL_SW_PARAMS: |
3318 | return snd_pcm_sw_params_user(substream, params: arg); |
3319 | case SNDRV_PCM_IOCTL_STATUS32: |
3320 | return snd_pcm_status_user32(substream, status: arg, ext: false); |
3321 | case SNDRV_PCM_IOCTL_STATUS_EXT32: |
3322 | return snd_pcm_status_user32(substream, status: arg, ext: true); |
3323 | case SNDRV_PCM_IOCTL_STATUS64: |
3324 | return snd_pcm_status_user64(substream, status: arg, ext: false); |
3325 | case SNDRV_PCM_IOCTL_STATUS_EXT64: |
3326 | return snd_pcm_status_user64(substream, status: arg, ext: true); |
3327 | case SNDRV_PCM_IOCTL_CHANNEL_INFO: |
3328 | return snd_pcm_channel_info_user(substream, info: arg); |
3329 | case SNDRV_PCM_IOCTL_PREPARE: |
3330 | return snd_pcm_prepare(substream, file); |
3331 | case SNDRV_PCM_IOCTL_RESET: |
3332 | return snd_pcm_reset(substream); |
3333 | case SNDRV_PCM_IOCTL_START: |
3334 | return snd_pcm_start_lock_irq(substream); |
3335 | case SNDRV_PCM_IOCTL_LINK: |
3336 | return snd_pcm_link(substream, fd: (int)(unsigned long) arg); |
3337 | case SNDRV_PCM_IOCTL_UNLINK: |
3338 | return snd_pcm_unlink(substream); |
3339 | case SNDRV_PCM_IOCTL_RESUME: |
3340 | return snd_pcm_resume(substream); |
3341 | case SNDRV_PCM_IOCTL_XRUN: |
3342 | return snd_pcm_xrun(substream); |
3343 | case SNDRV_PCM_IOCTL_HWSYNC: |
3344 | return snd_pcm_hwsync(substream); |
3345 | case SNDRV_PCM_IOCTL_DELAY: |
3346 | { |
3347 | snd_pcm_sframes_t delay = 0; |
3348 | snd_pcm_sframes_t __user *res = arg; |
3349 | int err; |
3350 | |
3351 | err = snd_pcm_delay(substream, delay: &delay); |
3352 | if (err) |
3353 | return err; |
3354 | if (put_user(delay, res)) |
3355 | return -EFAULT; |
3356 | return 0; |
3357 | } |
3358 | case __SNDRV_PCM_IOCTL_SYNC_PTR32: |
3359 | return snd_pcm_ioctl_sync_ptr_compat(substream, src: arg); |
3360 | case __SNDRV_PCM_IOCTL_SYNC_PTR64: |
3361 | return snd_pcm_sync_ptr(substream, sync_ptr: arg); |
3362 | #ifdef CONFIG_SND_SUPPORT_OLD_API |
3363 | case SNDRV_PCM_IOCTL_HW_REFINE_OLD: |
3364 | return snd_pcm_hw_refine_old_user(substream, oparams: arg); |
3365 | case SNDRV_PCM_IOCTL_HW_PARAMS_OLD: |
3366 | return snd_pcm_hw_params_old_user(substream, oparams: arg); |
3367 | #endif |
3368 | case SNDRV_PCM_IOCTL_DRAIN: |
3369 | return snd_pcm_drain(substream, file); |
3370 | case SNDRV_PCM_IOCTL_DROP: |
3371 | return snd_pcm_drop(substream); |
3372 | case SNDRV_PCM_IOCTL_PAUSE: |
3373 | return snd_pcm_pause_lock_irq(substream, push: (unsigned long)arg); |
3374 | case SNDRV_PCM_IOCTL_WRITEI_FRAMES: |
3375 | case SNDRV_PCM_IOCTL_READI_FRAMES: |
3376 | return snd_pcm_xferi_frames_ioctl(substream, xferi: arg); |
3377 | case SNDRV_PCM_IOCTL_WRITEN_FRAMES: |
3378 | case SNDRV_PCM_IOCTL_READN_FRAMES: |
3379 | return snd_pcm_xfern_frames_ioctl(substream, xfern: arg); |
3380 | case SNDRV_PCM_IOCTL_REWIND: |
3381 | return snd_pcm_rewind_ioctl(substream, frames: arg); |
3382 | case SNDRV_PCM_IOCTL_FORWARD: |
3383 | return snd_pcm_forward_ioctl(substream, frames: arg); |
3384 | } |
3385 | pcm_dbg(substream->pcm, "unknown ioctl = 0x%x\n" , cmd); |
3386 | return -ENOTTY; |
3387 | } |
3388 | |
3389 | static long snd_pcm_ioctl(struct file *file, unsigned int cmd, |
3390 | unsigned long arg) |
3391 | { |
3392 | struct snd_pcm_file *pcm_file; |
3393 | |
3394 | pcm_file = file->private_data; |
3395 | |
3396 | if (((cmd >> 8) & 0xff) != 'A') |
3397 | return -ENOTTY; |
3398 | |
3399 | return snd_pcm_common_ioctl(file, substream: pcm_file->substream, cmd, |
3400 | arg: (void __user *)arg); |
3401 | } |
3402 | |
3403 | /** |
3404 | * snd_pcm_kernel_ioctl - Execute PCM ioctl in the kernel-space |
3405 | * @substream: PCM substream |
3406 | * @cmd: IOCTL cmd |
3407 | * @arg: IOCTL argument |
3408 | * |
3409 | * The function is provided primarily for OSS layer and USB gadget drivers, |
3410 | * and it allows only the limited set of ioctls (hw_params, sw_params, |
3411 | * prepare, start, drain, drop, forward). |
3412 | * |
3413 | * Return: zero if successful, or a negative error code |
3414 | */ |
3415 | int snd_pcm_kernel_ioctl(struct snd_pcm_substream *substream, |
3416 | unsigned int cmd, void *arg) |
3417 | { |
3418 | snd_pcm_uframes_t *frames = arg; |
3419 | snd_pcm_sframes_t result; |
3420 | |
3421 | if (substream->runtime->state == SNDRV_PCM_STATE_DISCONNECTED) |
3422 | return -EBADFD; |
3423 | |
3424 | switch (cmd) { |
3425 | case SNDRV_PCM_IOCTL_FORWARD: |
3426 | { |
3427 | /* provided only for OSS; capture-only and no value returned */ |
3428 | if (substream->stream != SNDRV_PCM_STREAM_CAPTURE) |
3429 | return -EINVAL; |
3430 | result = snd_pcm_forward(substream, frames: *frames); |
3431 | return result < 0 ? result : 0; |
3432 | } |
3433 | case SNDRV_PCM_IOCTL_HW_PARAMS: |
3434 | return snd_pcm_hw_params(substream, params: arg); |
3435 | case SNDRV_PCM_IOCTL_SW_PARAMS: |
3436 | return snd_pcm_sw_params(substream, params: arg); |
3437 | case SNDRV_PCM_IOCTL_PREPARE: |
3438 | return snd_pcm_prepare(substream, NULL); |
3439 | case SNDRV_PCM_IOCTL_START: |
3440 | return snd_pcm_start_lock_irq(substream); |
3441 | case SNDRV_PCM_IOCTL_DRAIN: |
3442 | return snd_pcm_drain(substream, NULL); |
3443 | case SNDRV_PCM_IOCTL_DROP: |
3444 | return snd_pcm_drop(substream); |
3445 | case SNDRV_PCM_IOCTL_DELAY: |
3446 | return snd_pcm_delay(substream, delay: frames); |
3447 | default: |
3448 | return -EINVAL; |
3449 | } |
3450 | } |
3451 | EXPORT_SYMBOL(snd_pcm_kernel_ioctl); |
3452 | |
3453 | static ssize_t snd_pcm_read(struct file *file, char __user *buf, size_t count, |
3454 | loff_t * offset) |
3455 | { |
3456 | struct snd_pcm_file *pcm_file; |
3457 | struct snd_pcm_substream *substream; |
3458 | struct snd_pcm_runtime *runtime; |
3459 | snd_pcm_sframes_t result; |
3460 | |
3461 | pcm_file = file->private_data; |
3462 | substream = pcm_file->substream; |
3463 | if (PCM_RUNTIME_CHECK(substream)) |
3464 | return -ENXIO; |
3465 | runtime = substream->runtime; |
3466 | if (runtime->state == SNDRV_PCM_STATE_OPEN || |
3467 | runtime->state == SNDRV_PCM_STATE_DISCONNECTED) |
3468 | return -EBADFD; |
3469 | if (!frame_aligned(runtime, bytes: count)) |
3470 | return -EINVAL; |
3471 | count = bytes_to_frames(runtime, size: count); |
3472 | result = snd_pcm_lib_read(substream, buf, frames: count); |
3473 | if (result > 0) |
3474 | result = frames_to_bytes(runtime, size: result); |
3475 | return result; |
3476 | } |
3477 | |
3478 | static ssize_t snd_pcm_write(struct file *file, const char __user *buf, |
3479 | size_t count, loff_t * offset) |
3480 | { |
3481 | struct snd_pcm_file *pcm_file; |
3482 | struct snd_pcm_substream *substream; |
3483 | struct snd_pcm_runtime *runtime; |
3484 | snd_pcm_sframes_t result; |
3485 | |
3486 | pcm_file = file->private_data; |
3487 | substream = pcm_file->substream; |
3488 | if (PCM_RUNTIME_CHECK(substream)) |
3489 | return -ENXIO; |
3490 | runtime = substream->runtime; |
3491 | if (runtime->state == SNDRV_PCM_STATE_OPEN || |
3492 | runtime->state == SNDRV_PCM_STATE_DISCONNECTED) |
3493 | return -EBADFD; |
3494 | if (!frame_aligned(runtime, bytes: count)) |
3495 | return -EINVAL; |
3496 | count = bytes_to_frames(runtime, size: count); |
3497 | result = snd_pcm_lib_write(substream, buf, frames: count); |
3498 | if (result > 0) |
3499 | result = frames_to_bytes(runtime, size: result); |
3500 | return result; |
3501 | } |
3502 | |
3503 | static ssize_t snd_pcm_readv(struct kiocb *iocb, struct iov_iter *to) |
3504 | { |
3505 | struct snd_pcm_file *pcm_file; |
3506 | struct snd_pcm_substream *substream; |
3507 | struct snd_pcm_runtime *runtime; |
3508 | snd_pcm_sframes_t result; |
3509 | unsigned long i; |
3510 | void __user **bufs __free(kfree) = NULL; |
3511 | snd_pcm_uframes_t frames; |
3512 | const struct iovec *iov = iter_iov(iter: to); |
3513 | |
3514 | pcm_file = iocb->ki_filp->private_data; |
3515 | substream = pcm_file->substream; |
3516 | if (PCM_RUNTIME_CHECK(substream)) |
3517 | return -ENXIO; |
3518 | runtime = substream->runtime; |
3519 | if (runtime->state == SNDRV_PCM_STATE_OPEN || |
3520 | runtime->state == SNDRV_PCM_STATE_DISCONNECTED) |
3521 | return -EBADFD; |
3522 | if (!user_backed_iter(i: to)) |
3523 | return -EINVAL; |
3524 | if (to->nr_segs > 1024 || to->nr_segs != runtime->channels) |
3525 | return -EINVAL; |
3526 | if (!frame_aligned(runtime, bytes: iov->iov_len)) |
3527 | return -EINVAL; |
3528 | frames = bytes_to_samples(runtime, size: iov->iov_len); |
3529 | bufs = kmalloc_array(n: to->nr_segs, size: sizeof(void *), GFP_KERNEL); |
3530 | if (bufs == NULL) |
3531 | return -ENOMEM; |
3532 | for (i = 0; i < to->nr_segs; ++i) { |
3533 | bufs[i] = iov->iov_base; |
3534 | iov++; |
3535 | } |
3536 | result = snd_pcm_lib_readv(substream, bufs, frames); |
3537 | if (result > 0) |
3538 | result = frames_to_bytes(runtime, size: result); |
3539 | return result; |
3540 | } |
3541 | |
3542 | static ssize_t snd_pcm_writev(struct kiocb *iocb, struct iov_iter *from) |
3543 | { |
3544 | struct snd_pcm_file *pcm_file; |
3545 | struct snd_pcm_substream *substream; |
3546 | struct snd_pcm_runtime *runtime; |
3547 | snd_pcm_sframes_t result; |
3548 | unsigned long i; |
3549 | void __user **bufs __free(kfree) = NULL; |
3550 | snd_pcm_uframes_t frames; |
3551 | const struct iovec *iov = iter_iov(iter: from); |
3552 | |
3553 | pcm_file = iocb->ki_filp->private_data; |
3554 | substream = pcm_file->substream; |
3555 | if (PCM_RUNTIME_CHECK(substream)) |
3556 | return -ENXIO; |
3557 | runtime = substream->runtime; |
3558 | if (runtime->state == SNDRV_PCM_STATE_OPEN || |
3559 | runtime->state == SNDRV_PCM_STATE_DISCONNECTED) |
3560 | return -EBADFD; |
3561 | if (!user_backed_iter(i: from)) |
3562 | return -EINVAL; |
3563 | if (from->nr_segs > 128 || from->nr_segs != runtime->channels || |
3564 | !frame_aligned(runtime, bytes: iov->iov_len)) |
3565 | return -EINVAL; |
3566 | frames = bytes_to_samples(runtime, size: iov->iov_len); |
3567 | bufs = kmalloc_array(n: from->nr_segs, size: sizeof(void *), GFP_KERNEL); |
3568 | if (bufs == NULL) |
3569 | return -ENOMEM; |
3570 | for (i = 0; i < from->nr_segs; ++i) { |
3571 | bufs[i] = iov->iov_base; |
3572 | iov++; |
3573 | } |
3574 | result = snd_pcm_lib_writev(substream, bufs, frames); |
3575 | if (result > 0) |
3576 | result = frames_to_bytes(runtime, size: result); |
3577 | return result; |
3578 | } |
3579 | |
3580 | static __poll_t snd_pcm_poll(struct file *file, poll_table *wait) |
3581 | { |
3582 | struct snd_pcm_file *pcm_file; |
3583 | struct snd_pcm_substream *substream; |
3584 | struct snd_pcm_runtime *runtime; |
3585 | __poll_t mask, ok; |
3586 | snd_pcm_uframes_t avail; |
3587 | |
3588 | pcm_file = file->private_data; |
3589 | |
3590 | substream = pcm_file->substream; |
3591 | if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) |
3592 | ok = EPOLLOUT | EPOLLWRNORM; |
3593 | else |
3594 | ok = EPOLLIN | EPOLLRDNORM; |
3595 | if (PCM_RUNTIME_CHECK(substream)) |
3596 | return ok | EPOLLERR; |
3597 | |
3598 | runtime = substream->runtime; |
3599 | if (runtime->state == SNDRV_PCM_STATE_DISCONNECTED) |
3600 | return ok | EPOLLERR; |
3601 | |
3602 | poll_wait(filp: file, wait_address: &runtime->sleep, p: wait); |
3603 | |
3604 | mask = 0; |
3605 | guard(pcm_stream_lock_irq)(l: substream); |
3606 | avail = snd_pcm_avail(substream); |
3607 | switch (runtime->state) { |
3608 | case SNDRV_PCM_STATE_RUNNING: |
3609 | case SNDRV_PCM_STATE_PREPARED: |
3610 | case SNDRV_PCM_STATE_PAUSED: |
3611 | if (avail >= runtime->control->avail_min) |
3612 | mask = ok; |
3613 | break; |
3614 | case SNDRV_PCM_STATE_DRAINING: |
3615 | if (substream->stream == SNDRV_PCM_STREAM_CAPTURE) { |
3616 | mask = ok; |
3617 | if (!avail) |
3618 | mask |= EPOLLERR; |
3619 | } |
3620 | break; |
3621 | default: |
3622 | mask = ok | EPOLLERR; |
3623 | break; |
3624 | } |
3625 | return mask; |
3626 | } |
3627 | |
3628 | /* |
3629 | * mmap support |
3630 | */ |
3631 | |
3632 | /* |
3633 | * Only on coherent architectures, we can mmap the status and the control records |
3634 | * for effcient data transfer. On others, we have to use HWSYNC ioctl... |
3635 | */ |
3636 | #if defined(CONFIG_X86) || defined(CONFIG_PPC) || defined(CONFIG_ALPHA) |
3637 | /* |
3638 | * mmap status record |
3639 | */ |
3640 | static vm_fault_t snd_pcm_mmap_status_fault(struct vm_fault *vmf) |
3641 | { |
3642 | struct snd_pcm_substream *substream = vmf->vma->vm_private_data; |
3643 | struct snd_pcm_runtime *runtime; |
3644 | |
3645 | if (substream == NULL) |
3646 | return VM_FAULT_SIGBUS; |
3647 | runtime = substream->runtime; |
3648 | vmf->page = virt_to_page(runtime->status); |
3649 | get_page(page: vmf->page); |
3650 | return 0; |
3651 | } |
3652 | |
3653 | static const struct vm_operations_struct snd_pcm_vm_ops_status = |
3654 | { |
3655 | .fault = snd_pcm_mmap_status_fault, |
3656 | }; |
3657 | |
3658 | static int snd_pcm_mmap_status(struct snd_pcm_substream *substream, struct file *file, |
3659 | struct vm_area_struct *area) |
3660 | { |
3661 | long size; |
3662 | if (!(area->vm_flags & VM_READ)) |
3663 | return -EINVAL; |
3664 | size = area->vm_end - area->vm_start; |
3665 | if (size != PAGE_ALIGN(sizeof(struct snd_pcm_mmap_status))) |
3666 | return -EINVAL; |
3667 | area->vm_ops = &snd_pcm_vm_ops_status; |
3668 | area->vm_private_data = substream; |
3669 | vm_flags_mod(vma: area, VM_DONTEXPAND | VM_DONTDUMP, |
3670 | VM_WRITE | VM_MAYWRITE); |
3671 | |
3672 | return 0; |
3673 | } |
3674 | |
3675 | /* |
3676 | * mmap control record |
3677 | */ |
3678 | static vm_fault_t snd_pcm_mmap_control_fault(struct vm_fault *vmf) |
3679 | { |
3680 | struct snd_pcm_substream *substream = vmf->vma->vm_private_data; |
3681 | struct snd_pcm_runtime *runtime; |
3682 | |
3683 | if (substream == NULL) |
3684 | return VM_FAULT_SIGBUS; |
3685 | runtime = substream->runtime; |
3686 | vmf->page = virt_to_page(runtime->control); |
3687 | get_page(page: vmf->page); |
3688 | return 0; |
3689 | } |
3690 | |
3691 | static const struct vm_operations_struct snd_pcm_vm_ops_control = |
3692 | { |
3693 | .fault = snd_pcm_mmap_control_fault, |
3694 | }; |
3695 | |
3696 | static int snd_pcm_mmap_control(struct snd_pcm_substream *substream, struct file *file, |
3697 | struct vm_area_struct *area) |
3698 | { |
3699 | long size; |
3700 | if (!(area->vm_flags & VM_READ)) |
3701 | return -EINVAL; |
3702 | size = area->vm_end - area->vm_start; |
3703 | if (size != PAGE_ALIGN(sizeof(struct snd_pcm_mmap_control))) |
3704 | return -EINVAL; |
3705 | area->vm_ops = &snd_pcm_vm_ops_control; |
3706 | area->vm_private_data = substream; |
3707 | vm_flags_set(vma: area, VM_DONTEXPAND | VM_DONTDUMP); |
3708 | return 0; |
3709 | } |
3710 | |
3711 | static bool pcm_status_mmap_allowed(struct snd_pcm_file *pcm_file) |
3712 | { |
3713 | /* If drivers require the explicit sync (typically for non-coherent |
3714 | * pages), we have to disable the mmap of status and control data |
3715 | * to enforce the control via SYNC_PTR ioctl. |
3716 | */ |
3717 | if (pcm_file->substream->runtime->hw.info & SNDRV_PCM_INFO_EXPLICIT_SYNC) |
3718 | return false; |
3719 | /* See pcm_control_mmap_allowed() below. |
3720 | * Since older alsa-lib requires both status and control mmaps to be |
3721 | * coupled, we have to disable the status mmap for old alsa-lib, too. |
3722 | */ |
3723 | if (pcm_file->user_pversion < SNDRV_PROTOCOL_VERSION(2, 0, 14) && |
3724 | (pcm_file->substream->runtime->hw.info & SNDRV_PCM_INFO_SYNC_APPLPTR)) |
3725 | return false; |
3726 | return true; |
3727 | } |
3728 | |
3729 | static bool pcm_control_mmap_allowed(struct snd_pcm_file *pcm_file) |
3730 | { |
3731 | if (pcm_file->no_compat_mmap) |
3732 | return false; |
3733 | /* see above */ |
3734 | if (pcm_file->substream->runtime->hw.info & SNDRV_PCM_INFO_EXPLICIT_SYNC) |
3735 | return false; |
3736 | /* Disallow the control mmap when SYNC_APPLPTR flag is set; |
3737 | * it enforces the user-space to fall back to snd_pcm_sync_ptr(), |
3738 | * thus it effectively assures the manual update of appl_ptr. |
3739 | */ |
3740 | if (pcm_file->substream->runtime->hw.info & SNDRV_PCM_INFO_SYNC_APPLPTR) |
3741 | return false; |
3742 | return true; |
3743 | } |
3744 | |
3745 | #else /* ! coherent mmap */ |
3746 | /* |
3747 | * don't support mmap for status and control records. |
3748 | */ |
3749 | #define pcm_status_mmap_allowed(pcm_file) false |
3750 | #define pcm_control_mmap_allowed(pcm_file) false |
3751 | |
3752 | static int snd_pcm_mmap_status(struct snd_pcm_substream *substream, struct file *file, |
3753 | struct vm_area_struct *area) |
3754 | { |
3755 | return -ENXIO; |
3756 | } |
3757 | static int snd_pcm_mmap_control(struct snd_pcm_substream *substream, struct file *file, |
3758 | struct vm_area_struct *area) |
3759 | { |
3760 | return -ENXIO; |
3761 | } |
3762 | #endif /* coherent mmap */ |
3763 | |
3764 | /* |
3765 | * fault callback for mmapping a RAM page |
3766 | */ |
3767 | static vm_fault_t snd_pcm_mmap_data_fault(struct vm_fault *vmf) |
3768 | { |
3769 | struct snd_pcm_substream *substream = vmf->vma->vm_private_data; |
3770 | struct snd_pcm_runtime *runtime; |
3771 | unsigned long offset; |
3772 | struct page * page; |
3773 | size_t dma_bytes; |
3774 | |
3775 | if (substream == NULL) |
3776 | return VM_FAULT_SIGBUS; |
3777 | runtime = substream->runtime; |
3778 | offset = vmf->pgoff << PAGE_SHIFT; |
3779 | dma_bytes = PAGE_ALIGN(runtime->dma_bytes); |
3780 | if (offset > dma_bytes - PAGE_SIZE) |
3781 | return VM_FAULT_SIGBUS; |
3782 | if (substream->ops->page) |
3783 | page = substream->ops->page(substream, offset); |
3784 | else if (!snd_pcm_get_dma_buf(substream)) |
3785 | page = virt_to_page(runtime->dma_area + offset); |
3786 | else |
3787 | page = snd_sgbuf_get_page(snd_pcm_get_dma_buf(substream), offset); |
3788 | if (!page) |
3789 | return VM_FAULT_SIGBUS; |
3790 | get_page(page); |
3791 | vmf->page = page; |
3792 | return 0; |
3793 | } |
3794 | |
3795 | static const struct vm_operations_struct snd_pcm_vm_ops_data = { |
3796 | .open = snd_pcm_mmap_data_open, |
3797 | .close = snd_pcm_mmap_data_close, |
3798 | }; |
3799 | |
3800 | static const struct vm_operations_struct snd_pcm_vm_ops_data_fault = { |
3801 | .open = snd_pcm_mmap_data_open, |
3802 | .close = snd_pcm_mmap_data_close, |
3803 | .fault = snd_pcm_mmap_data_fault, |
3804 | }; |
3805 | |
3806 | /* |
3807 | * mmap the DMA buffer on RAM |
3808 | */ |
3809 | |
3810 | /** |
3811 | * snd_pcm_lib_default_mmap - Default PCM data mmap function |
3812 | * @substream: PCM substream |
3813 | * @area: VMA |
3814 | * |
3815 | * This is the default mmap handler for PCM data. When mmap pcm_ops is NULL, |
3816 | * this function is invoked implicitly. |
3817 | * |
3818 | * Return: zero if successful, or a negative error code |
3819 | */ |
3820 | int snd_pcm_lib_default_mmap(struct snd_pcm_substream *substream, |
3821 | struct vm_area_struct *area) |
3822 | { |
3823 | vm_flags_set(vma: area, VM_DONTEXPAND | VM_DONTDUMP); |
3824 | if (!substream->ops->page && |
3825 | !snd_dma_buffer_mmap(snd_pcm_get_dma_buf(substream), area)) |
3826 | return 0; |
3827 | /* mmap with fault handler */ |
3828 | area->vm_ops = &snd_pcm_vm_ops_data_fault; |
3829 | return 0; |
3830 | } |
3831 | EXPORT_SYMBOL_GPL(snd_pcm_lib_default_mmap); |
3832 | |
3833 | /* |
3834 | * mmap the DMA buffer on I/O memory area |
3835 | */ |
3836 | #if SNDRV_PCM_INFO_MMAP_IOMEM |
3837 | /** |
3838 | * snd_pcm_lib_mmap_iomem - Default PCM data mmap function for I/O mem |
3839 | * @substream: PCM substream |
3840 | * @area: VMA |
3841 | * |
3842 | * When your hardware uses the iomapped pages as the hardware buffer and |
3843 | * wants to mmap it, pass this function as mmap pcm_ops. Note that this |
3844 | * is supposed to work only on limited architectures. |
3845 | * |
3846 | * Return: zero if successful, or a negative error code |
3847 | */ |
3848 | int snd_pcm_lib_mmap_iomem(struct snd_pcm_substream *substream, |
3849 | struct vm_area_struct *area) |
3850 | { |
3851 | struct snd_pcm_runtime *runtime = substream->runtime; |
3852 | |
3853 | area->vm_page_prot = pgprot_noncached(area->vm_page_prot); |
3854 | return vm_iomap_memory(vma: area, start: runtime->dma_addr, len: runtime->dma_bytes); |
3855 | } |
3856 | EXPORT_SYMBOL(snd_pcm_lib_mmap_iomem); |
3857 | #endif /* SNDRV_PCM_INFO_MMAP */ |
3858 | |
3859 | /* |
3860 | * mmap DMA buffer |
3861 | */ |
3862 | int snd_pcm_mmap_data(struct snd_pcm_substream *substream, struct file *file, |
3863 | struct vm_area_struct *area) |
3864 | { |
3865 | struct snd_pcm_runtime *runtime; |
3866 | long size; |
3867 | unsigned long offset; |
3868 | size_t dma_bytes; |
3869 | int err; |
3870 | |
3871 | if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) { |
3872 | if (!(area->vm_flags & (VM_WRITE|VM_READ))) |
3873 | return -EINVAL; |
3874 | } else { |
3875 | if (!(area->vm_flags & VM_READ)) |
3876 | return -EINVAL; |
3877 | } |
3878 | runtime = substream->runtime; |
3879 | if (runtime->state == SNDRV_PCM_STATE_OPEN) |
3880 | return -EBADFD; |
3881 | if (!(runtime->info & SNDRV_PCM_INFO_MMAP)) |
3882 | return -ENXIO; |
3883 | if (runtime->access == SNDRV_PCM_ACCESS_RW_INTERLEAVED || |
3884 | runtime->access == SNDRV_PCM_ACCESS_RW_NONINTERLEAVED) |
3885 | return -EINVAL; |
3886 | size = area->vm_end - area->vm_start; |
3887 | offset = area->vm_pgoff << PAGE_SHIFT; |
3888 | dma_bytes = PAGE_ALIGN(runtime->dma_bytes); |
3889 | if ((size_t)size > dma_bytes) |
3890 | return -EINVAL; |
3891 | if (offset > dma_bytes - size) |
3892 | return -EINVAL; |
3893 | |
3894 | area->vm_ops = &snd_pcm_vm_ops_data; |
3895 | area->vm_private_data = substream; |
3896 | if (substream->ops->mmap) |
3897 | err = substream->ops->mmap(substream, area); |
3898 | else |
3899 | err = snd_pcm_lib_default_mmap(substream, area); |
3900 | if (!err) |
3901 | atomic_inc(v: &substream->mmap_count); |
3902 | return err; |
3903 | } |
3904 | EXPORT_SYMBOL(snd_pcm_mmap_data); |
3905 | |
3906 | static int snd_pcm_mmap(struct file *file, struct vm_area_struct *area) |
3907 | { |
3908 | struct snd_pcm_file * pcm_file; |
3909 | struct snd_pcm_substream *substream; |
3910 | unsigned long offset; |
3911 | |
3912 | pcm_file = file->private_data; |
3913 | substream = pcm_file->substream; |
3914 | if (PCM_RUNTIME_CHECK(substream)) |
3915 | return -ENXIO; |
3916 | if (substream->runtime->state == SNDRV_PCM_STATE_DISCONNECTED) |
3917 | return -EBADFD; |
3918 | |
3919 | offset = area->vm_pgoff << PAGE_SHIFT; |
3920 | switch (offset) { |
3921 | case SNDRV_PCM_MMAP_OFFSET_STATUS_OLD: |
3922 | if (pcm_file->no_compat_mmap || !IS_ENABLED(CONFIG_64BIT)) |
3923 | return -ENXIO; |
3924 | fallthrough; |
3925 | case SNDRV_PCM_MMAP_OFFSET_STATUS_NEW: |
3926 | if (!pcm_status_mmap_allowed(pcm_file)) |
3927 | return -ENXIO; |
3928 | return snd_pcm_mmap_status(substream, file, area); |
3929 | case SNDRV_PCM_MMAP_OFFSET_CONTROL_OLD: |
3930 | if (pcm_file->no_compat_mmap || !IS_ENABLED(CONFIG_64BIT)) |
3931 | return -ENXIO; |
3932 | fallthrough; |
3933 | case SNDRV_PCM_MMAP_OFFSET_CONTROL_NEW: |
3934 | if (!pcm_control_mmap_allowed(pcm_file)) |
3935 | return -ENXIO; |
3936 | return snd_pcm_mmap_control(substream, file, area); |
3937 | default: |
3938 | return snd_pcm_mmap_data(substream, file, area); |
3939 | } |
3940 | return 0; |
3941 | } |
3942 | |
3943 | static int snd_pcm_fasync(int fd, struct file * file, int on) |
3944 | { |
3945 | struct snd_pcm_file * pcm_file; |
3946 | struct snd_pcm_substream *substream; |
3947 | struct snd_pcm_runtime *runtime; |
3948 | |
3949 | pcm_file = file->private_data; |
3950 | substream = pcm_file->substream; |
3951 | if (PCM_RUNTIME_CHECK(substream)) |
3952 | return -ENXIO; |
3953 | runtime = substream->runtime; |
3954 | if (runtime->state == SNDRV_PCM_STATE_DISCONNECTED) |
3955 | return -EBADFD; |
3956 | return snd_fasync_helper(fd, file, on, fasyncp: &runtime->fasync); |
3957 | } |
3958 | |
3959 | /* |
3960 | * ioctl32 compat |
3961 | */ |
3962 | #ifdef CONFIG_COMPAT |
3963 | #include "pcm_compat.c" |
3964 | #else |
3965 | #define snd_pcm_ioctl_compat NULL |
3966 | #endif |
3967 | |
3968 | /* |
3969 | * To be removed helpers to keep binary compatibility |
3970 | */ |
3971 | |
3972 | #ifdef CONFIG_SND_SUPPORT_OLD_API |
3973 | #define __OLD_TO_NEW_MASK(x) ((x&7)|((x&0x07fffff8)<<5)) |
3974 | #define __NEW_TO_OLD_MASK(x) ((x&7)|((x&0xffffff00)>>5)) |
3975 | |
3976 | static void snd_pcm_hw_convert_from_old_params(struct snd_pcm_hw_params *params, |
3977 | struct snd_pcm_hw_params_old *oparams) |
3978 | { |
3979 | unsigned int i; |
3980 | |
3981 | memset(params, 0, sizeof(*params)); |
3982 | params->flags = oparams->flags; |
3983 | for (i = 0; i < ARRAY_SIZE(oparams->masks); i++) |
3984 | params->masks[i].bits[0] = oparams->masks[i]; |
3985 | memcpy(params->intervals, oparams->intervals, sizeof(oparams->intervals)); |
3986 | params->rmask = __OLD_TO_NEW_MASK(oparams->rmask); |
3987 | params->cmask = __OLD_TO_NEW_MASK(oparams->cmask); |
3988 | params->info = oparams->info; |
3989 | params->msbits = oparams->msbits; |
3990 | params->rate_num = oparams->rate_num; |
3991 | params->rate_den = oparams->rate_den; |
3992 | params->fifo_size = oparams->fifo_size; |
3993 | } |
3994 | |
3995 | static void snd_pcm_hw_convert_to_old_params(struct snd_pcm_hw_params_old *oparams, |
3996 | struct snd_pcm_hw_params *params) |
3997 | { |
3998 | unsigned int i; |
3999 | |
4000 | memset(oparams, 0, sizeof(*oparams)); |
4001 | oparams->flags = params->flags; |
4002 | for (i = 0; i < ARRAY_SIZE(oparams->masks); i++) |
4003 | oparams->masks[i] = params->masks[i].bits[0]; |
4004 | memcpy(oparams->intervals, params->intervals, sizeof(oparams->intervals)); |
4005 | oparams->rmask = __NEW_TO_OLD_MASK(params->rmask); |
4006 | oparams->cmask = __NEW_TO_OLD_MASK(params->cmask); |
4007 | oparams->info = params->info; |
4008 | oparams->msbits = params->msbits; |
4009 | oparams->rate_num = params->rate_num; |
4010 | oparams->rate_den = params->rate_den; |
4011 | oparams->fifo_size = params->fifo_size; |
4012 | } |
4013 | |
4014 | static int snd_pcm_hw_refine_old_user(struct snd_pcm_substream *substream, |
4015 | struct snd_pcm_hw_params_old __user * _oparams) |
4016 | { |
4017 | struct snd_pcm_hw_params *params __free(kfree) = NULL; |
4018 | struct snd_pcm_hw_params_old *oparams __free(kfree) = NULL; |
4019 | int err; |
4020 | |
4021 | params = kmalloc(size: sizeof(*params), GFP_KERNEL); |
4022 | if (!params) |
4023 | return -ENOMEM; |
4024 | |
4025 | oparams = memdup_user(_oparams, sizeof(*oparams)); |
4026 | if (IS_ERR(ptr: oparams)) |
4027 | return PTR_ERR(no_free_ptr(oparams)); |
4028 | snd_pcm_hw_convert_from_old_params(params, oparams); |
4029 | err = snd_pcm_hw_refine(substream, params); |
4030 | if (err < 0) |
4031 | return err; |
4032 | |
4033 | err = fixup_unreferenced_params(substream, params); |
4034 | if (err < 0) |
4035 | return err; |
4036 | |
4037 | snd_pcm_hw_convert_to_old_params(oparams, params); |
4038 | if (copy_to_user(to: _oparams, from: oparams, n: sizeof(*oparams))) |
4039 | return -EFAULT; |
4040 | return 0; |
4041 | } |
4042 | |
4043 | static int snd_pcm_hw_params_old_user(struct snd_pcm_substream *substream, |
4044 | struct snd_pcm_hw_params_old __user * _oparams) |
4045 | { |
4046 | struct snd_pcm_hw_params *params __free(kfree) = NULL; |
4047 | struct snd_pcm_hw_params_old *oparams __free(kfree) = NULL; |
4048 | int err; |
4049 | |
4050 | params = kmalloc(size: sizeof(*params), GFP_KERNEL); |
4051 | if (!params) |
4052 | return -ENOMEM; |
4053 | |
4054 | oparams = memdup_user(_oparams, sizeof(*oparams)); |
4055 | if (IS_ERR(ptr: oparams)) |
4056 | return PTR_ERR(no_free_ptr(oparams)); |
4057 | |
4058 | snd_pcm_hw_convert_from_old_params(params, oparams); |
4059 | err = snd_pcm_hw_params(substream, params); |
4060 | if (err < 0) |
4061 | return err; |
4062 | |
4063 | snd_pcm_hw_convert_to_old_params(oparams, params); |
4064 | if (copy_to_user(to: _oparams, from: oparams, n: sizeof(*oparams))) |
4065 | return -EFAULT; |
4066 | return 0; |
4067 | } |
4068 | #endif /* CONFIG_SND_SUPPORT_OLD_API */ |
4069 | |
4070 | #ifndef CONFIG_MMU |
4071 | static unsigned long snd_pcm_get_unmapped_area(struct file *file, |
4072 | unsigned long addr, |
4073 | unsigned long len, |
4074 | unsigned long pgoff, |
4075 | unsigned long flags) |
4076 | { |
4077 | struct snd_pcm_file *pcm_file = file->private_data; |
4078 | struct snd_pcm_substream *substream = pcm_file->substream; |
4079 | struct snd_pcm_runtime *runtime = substream->runtime; |
4080 | unsigned long offset = pgoff << PAGE_SHIFT; |
4081 | |
4082 | switch (offset) { |
4083 | case SNDRV_PCM_MMAP_OFFSET_STATUS_NEW: |
4084 | return (unsigned long)runtime->status; |
4085 | case SNDRV_PCM_MMAP_OFFSET_CONTROL_NEW: |
4086 | return (unsigned long)runtime->control; |
4087 | default: |
4088 | return (unsigned long)runtime->dma_area + offset; |
4089 | } |
4090 | } |
4091 | #else |
4092 | # define snd_pcm_get_unmapped_area NULL |
4093 | #endif |
4094 | |
4095 | /* |
4096 | * Register section |
4097 | */ |
4098 | |
4099 | const struct file_operations snd_pcm_f_ops[2] = { |
4100 | { |
4101 | .owner = THIS_MODULE, |
4102 | .write = snd_pcm_write, |
4103 | .write_iter = snd_pcm_writev, |
4104 | .open = snd_pcm_playback_open, |
4105 | .release = snd_pcm_release, |
4106 | .llseek = no_llseek, |
4107 | .poll = snd_pcm_poll, |
4108 | .unlocked_ioctl = snd_pcm_ioctl, |
4109 | .compat_ioctl = snd_pcm_ioctl_compat, |
4110 | .mmap = snd_pcm_mmap, |
4111 | .fasync = snd_pcm_fasync, |
4112 | .get_unmapped_area = snd_pcm_get_unmapped_area, |
4113 | }, |
4114 | { |
4115 | .owner = THIS_MODULE, |
4116 | .read = snd_pcm_read, |
4117 | .read_iter = snd_pcm_readv, |
4118 | .open = snd_pcm_capture_open, |
4119 | .release = snd_pcm_release, |
4120 | .llseek = no_llseek, |
4121 | .poll = snd_pcm_poll, |
4122 | .unlocked_ioctl = snd_pcm_ioctl, |
4123 | .compat_ioctl = snd_pcm_ioctl_compat, |
4124 | .mmap = snd_pcm_mmap, |
4125 | .fasync = snd_pcm_fasync, |
4126 | .get_unmapped_area = snd_pcm_get_unmapped_area, |
4127 | } |
4128 | }; |
4129 | |