1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * ISHTP client logic
4 *
5 * Copyright (c) 2003-2016, Intel Corporation.
6 */
7
8#include <linux/slab.h>
9#include <linux/sched.h>
10#include <linux/wait.h>
11#include <linux/delay.h>
12#include <linux/dma-mapping.h>
13#include <asm/cacheflush.h>
14#include "hbm.h"
15#include "client.h"
16
17int ishtp_cl_get_tx_free_buffer_size(struct ishtp_cl *cl)
18{
19 unsigned long tx_free_flags;
20 int size;
21
22 spin_lock_irqsave(&cl->tx_free_list_spinlock, tx_free_flags);
23 size = cl->tx_ring_free_size * cl->device->fw_client->props.max_msg_length;
24 spin_unlock_irqrestore(lock: &cl->tx_free_list_spinlock, flags: tx_free_flags);
25
26 return size;
27}
28EXPORT_SYMBOL(ishtp_cl_get_tx_free_buffer_size);
29
30int ishtp_cl_get_tx_free_rings(struct ishtp_cl *cl)
31{
32 return cl->tx_ring_free_size;
33}
34EXPORT_SYMBOL(ishtp_cl_get_tx_free_rings);
35
36/**
37 * ishtp_read_list_flush() - Flush read queue
38 * @cl: ishtp client instance
39 *
40 * Used to remove all entries from read queue for a client
41 */
42static void ishtp_read_list_flush(struct ishtp_cl *cl)
43{
44 struct ishtp_cl_rb *rb;
45 struct ishtp_cl_rb *next;
46 unsigned long flags;
47
48 spin_lock_irqsave(&cl->dev->read_list_spinlock, flags);
49 list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list)
50 if (rb->cl && ishtp_cl_cmp_id(cl1: cl, cl2: rb->cl)) {
51 list_del(entry: &rb->list);
52 spin_lock(lock: &cl->free_list_spinlock);
53 list_add_tail(new: &rb->list, head: &cl->free_rb_list.list);
54 spin_unlock(lock: &cl->free_list_spinlock);
55 }
56 spin_unlock_irqrestore(lock: &cl->dev->read_list_spinlock, flags);
57}
58
59/**
60 * ishtp_cl_flush_queues() - Flush all queues for a client
61 * @cl: ishtp client instance
62 *
63 * Used to remove all queues for a client. This is called when a client device
64 * needs reset due to error, S3 resume or during module removal
65 *
66 * Return: 0 on success else -EINVAL if device is NULL
67 */
68int ishtp_cl_flush_queues(struct ishtp_cl *cl)
69{
70 if (WARN_ON(!cl || !cl->dev))
71 return -EINVAL;
72
73 ishtp_read_list_flush(cl);
74
75 return 0;
76}
77EXPORT_SYMBOL(ishtp_cl_flush_queues);
78
79/**
80 * ishtp_cl_init() - Initialize all fields of a client device
81 * @cl: ishtp client instance
82 * @dev: ishtp device
83 *
84 * Initializes a client device fields: Init spinlocks, init queues etc.
85 * This function is called during new client creation
86 */
87static void ishtp_cl_init(struct ishtp_cl *cl, struct ishtp_device *dev)
88{
89 memset(cl, 0, sizeof(struct ishtp_cl));
90 init_waitqueue_head(&cl->wait_ctrl_res);
91 spin_lock_init(&cl->free_list_spinlock);
92 spin_lock_init(&cl->in_process_spinlock);
93 spin_lock_init(&cl->tx_list_spinlock);
94 spin_lock_init(&cl->tx_free_list_spinlock);
95 spin_lock_init(&cl->fc_spinlock);
96 INIT_LIST_HEAD(list: &cl->link);
97 cl->dev = dev;
98
99 INIT_LIST_HEAD(list: &cl->free_rb_list.list);
100 INIT_LIST_HEAD(list: &cl->tx_list.list);
101 INIT_LIST_HEAD(list: &cl->tx_free_list.list);
102 INIT_LIST_HEAD(list: &cl->in_process_list.list);
103
104 cl->rx_ring_size = CL_DEF_RX_RING_SIZE;
105 cl->tx_ring_size = CL_DEF_TX_RING_SIZE;
106 cl->tx_ring_free_size = cl->tx_ring_size;
107
108 /* dma */
109 cl->last_tx_path = CL_TX_PATH_IPC;
110 cl->last_dma_acked = 1;
111 cl->last_dma_addr = NULL;
112 cl->last_ipc_acked = 1;
113}
114
115/**
116 * ishtp_cl_allocate() - allocates client structure and sets it up.
117 * @cl_device: ishtp client device
118 *
119 * Allocate memory for new client device and call to initialize each field.
120 *
121 * Return: The allocated client instance or NULL on failure
122 */
123struct ishtp_cl *ishtp_cl_allocate(struct ishtp_cl_device *cl_device)
124{
125 struct ishtp_cl *cl;
126
127 cl = kmalloc(size: sizeof(struct ishtp_cl), GFP_KERNEL);
128 if (!cl)
129 return NULL;
130
131 ishtp_cl_init(cl, dev: cl_device->ishtp_dev);
132 return cl;
133}
134EXPORT_SYMBOL(ishtp_cl_allocate);
135
136/**
137 * ishtp_cl_free() - Frees a client device
138 * @cl: client device instance
139 *
140 * Frees a client device
141 */
142void ishtp_cl_free(struct ishtp_cl *cl)
143{
144 struct ishtp_device *dev;
145 unsigned long flags;
146
147 if (!cl)
148 return;
149
150 dev = cl->dev;
151 if (!dev)
152 return;
153
154 spin_lock_irqsave(&dev->cl_list_lock, flags);
155 ishtp_cl_free_rx_ring(cl);
156 ishtp_cl_free_tx_ring(cl);
157 kfree(objp: cl);
158 spin_unlock_irqrestore(lock: &dev->cl_list_lock, flags);
159}
160EXPORT_SYMBOL(ishtp_cl_free);
161
162/**
163 * ishtp_cl_link() - Reserve a host id and link the client instance
164 * @cl: client device instance
165 *
166 * This allocates a single bit in the hostmap. This function will make sure
167 * that not many client sessions are opened at the same time. Once allocated
168 * the client device instance is added to the ishtp device in the current
169 * client list
170 *
171 * Return: 0 or error code on failure
172 */
173int ishtp_cl_link(struct ishtp_cl *cl)
174{
175 struct ishtp_device *dev;
176 unsigned long flags, flags_cl;
177 int id, ret = 0;
178
179 if (WARN_ON(!cl || !cl->dev))
180 return -EINVAL;
181
182 dev = cl->dev;
183
184 spin_lock_irqsave(&dev->device_lock, flags);
185
186 if (dev->open_handle_count >= ISHTP_MAX_OPEN_HANDLE_COUNT) {
187 ret = -EMFILE;
188 goto unlock_dev;
189 }
190
191 id = find_first_zero_bit(addr: dev->host_clients_map, ISHTP_CLIENTS_MAX);
192
193 if (id >= ISHTP_CLIENTS_MAX) {
194 spin_unlock_irqrestore(lock: &dev->device_lock, flags);
195 dev_err(&cl->device->dev, "id exceeded %d", ISHTP_CLIENTS_MAX);
196 return -ENOENT;
197 }
198
199 dev->open_handle_count++;
200 cl->host_client_id = id;
201 spin_lock_irqsave(&dev->cl_list_lock, flags_cl);
202 if (dev->dev_state != ISHTP_DEV_ENABLED) {
203 ret = -ENODEV;
204 goto unlock_cl;
205 }
206 list_add_tail(new: &cl->link, head: &dev->cl_list);
207 set_bit(nr: id, addr: dev->host_clients_map);
208 cl->state = ISHTP_CL_INITIALIZING;
209
210unlock_cl:
211 spin_unlock_irqrestore(lock: &dev->cl_list_lock, flags: flags_cl);
212unlock_dev:
213 spin_unlock_irqrestore(lock: &dev->device_lock, flags);
214 return ret;
215}
216EXPORT_SYMBOL(ishtp_cl_link);
217
218/**
219 * ishtp_cl_unlink() - remove fw_cl from the client device list
220 * @cl: client device instance
221 *
222 * Remove a previously linked device to a ishtp device
223 */
224void ishtp_cl_unlink(struct ishtp_cl *cl)
225{
226 struct ishtp_device *dev;
227 struct ishtp_cl *pos;
228 unsigned long flags;
229
230 /* don't shout on error exit path */
231 if (!cl || !cl->dev)
232 return;
233
234 dev = cl->dev;
235
236 spin_lock_irqsave(&dev->device_lock, flags);
237 if (dev->open_handle_count > 0) {
238 clear_bit(nr: cl->host_client_id, addr: dev->host_clients_map);
239 dev->open_handle_count--;
240 }
241 spin_unlock_irqrestore(lock: &dev->device_lock, flags);
242
243 /*
244 * This checks that 'cl' is actually linked into device's structure,
245 * before attempting 'list_del'
246 */
247 spin_lock_irqsave(&dev->cl_list_lock, flags);
248 list_for_each_entry(pos, &dev->cl_list, link)
249 if (cl->host_client_id == pos->host_client_id) {
250 list_del_init(entry: &pos->link);
251 break;
252 }
253 spin_unlock_irqrestore(lock: &dev->cl_list_lock, flags);
254}
255EXPORT_SYMBOL(ishtp_cl_unlink);
256
257/**
258 * ishtp_cl_disconnect() - Send disconnect request to firmware
259 * @cl: client device instance
260 *
261 * Send a disconnect request for a client to firmware.
262 *
263 * Return: 0 if successful disconnect response from the firmware or error
264 * code on failure
265 */
266int ishtp_cl_disconnect(struct ishtp_cl *cl)
267{
268 struct ishtp_device *dev;
269
270 if (WARN_ON(!cl || !cl->dev))
271 return -ENODEV;
272
273 dev = cl->dev;
274
275 dev->print_log(dev, "%s() state %d\n", __func__, cl->state);
276
277 if (cl->state != ISHTP_CL_DISCONNECTING) {
278 dev->print_log(dev, "%s() Disconnect in progress\n", __func__);
279 return 0;
280 }
281
282 if (ishtp_hbm_cl_disconnect_req(dev, cl)) {
283 dev->print_log(dev, "%s() Failed to disconnect\n", __func__);
284 dev_err(&cl->device->dev, "failed to disconnect.\n");
285 return -ENODEV;
286 }
287
288 wait_event_interruptible_timeout(cl->wait_ctrl_res,
289 (dev->dev_state != ISHTP_DEV_ENABLED ||
290 cl->state == ISHTP_CL_DISCONNECTED),
291 ishtp_secs_to_jiffies(ISHTP_CL_CONNECT_TIMEOUT));
292
293 /*
294 * If FW reset arrived, this will happen. Don't check cl->,
295 * as 'cl' may be freed already
296 */
297 if (dev->dev_state != ISHTP_DEV_ENABLED) {
298 dev->print_log(dev, "%s() dev_state != ISHTP_DEV_ENABLED\n",
299 __func__);
300 return -ENODEV;
301 }
302
303 if (cl->state == ISHTP_CL_DISCONNECTED) {
304 dev->print_log(dev, "%s() successful\n", __func__);
305 return 0;
306 }
307
308 return -ENODEV;
309}
310EXPORT_SYMBOL(ishtp_cl_disconnect);
311
312/**
313 * ishtp_cl_is_other_connecting() - Check other client is connecting
314 * @cl: client device instance
315 *
316 * Checks if other client with the same fw client id is connecting
317 *
318 * Return: true if other client is connected else false
319 */
320static bool ishtp_cl_is_other_connecting(struct ishtp_cl *cl)
321{
322 struct ishtp_device *dev;
323 struct ishtp_cl *pos;
324 unsigned long flags;
325
326 if (WARN_ON(!cl || !cl->dev))
327 return false;
328
329 dev = cl->dev;
330 spin_lock_irqsave(&dev->cl_list_lock, flags);
331 list_for_each_entry(pos, &dev->cl_list, link) {
332 if ((pos->state == ISHTP_CL_CONNECTING) && (pos != cl) &&
333 cl->fw_client_id == pos->fw_client_id) {
334 spin_unlock_irqrestore(lock: &dev->cl_list_lock, flags);
335 return true;
336 }
337 }
338 spin_unlock_irqrestore(lock: &dev->cl_list_lock, flags);
339
340 return false;
341}
342
343/**
344 * ishtp_cl_connect_to_fw() - Send connect request to firmware
345 * @cl: client device instance
346 *
347 * Send a connect request to the firmware and wait for firmware response.
348 * If there is successful connection response from the firmware, change
349 * client state to ISHTP_CL_CONNECTED, and bind client to related
350 * firmware client_id.
351 *
352 * Return: 0 for success and error code on failure
353 */
354static int ishtp_cl_connect_to_fw(struct ishtp_cl *cl)
355{
356 struct ishtp_device *dev;
357 int rets;
358
359 if (WARN_ON(!cl || !cl->dev))
360 return -ENODEV;
361
362 dev = cl->dev;
363
364 if (ishtp_cl_is_other_connecting(cl)) {
365 dev->print_log(dev, "%s() Busy\n", __func__);
366 return -EBUSY;
367 }
368
369 if (ishtp_hbm_cl_connect_req(dev, cl)) {
370 dev->print_log(dev, "%s() HBM connect req fail\n", __func__);
371 return -ENODEV;
372 }
373
374 rets = wait_event_interruptible_timeout(cl->wait_ctrl_res,
375 (dev->dev_state == ISHTP_DEV_ENABLED &&
376 (cl->state == ISHTP_CL_CONNECTED ||
377 cl->state == ISHTP_CL_DISCONNECTED)),
378 ishtp_secs_to_jiffies(
379 ISHTP_CL_CONNECT_TIMEOUT));
380 /*
381 * If FW reset arrived, this will happen. Don't check cl->,
382 * as 'cl' may be freed already
383 */
384 if (dev->dev_state != ISHTP_DEV_ENABLED) {
385 dev->print_log(dev, "%s() dev_state != ISHTP_DEV_ENABLED\n",
386 __func__);
387 return -EFAULT;
388 }
389
390 if (cl->state != ISHTP_CL_CONNECTED) {
391 dev->print_log(dev, "%s() state != ISHTP_CL_CONNECTED\n",
392 __func__);
393 return -EFAULT;
394 }
395
396 rets = cl->status;
397 if (rets) {
398 dev->print_log(dev, "%s() Invalid status\n", __func__);
399 return rets;
400 }
401
402 rets = ishtp_cl_device_bind(cl);
403 if (rets) {
404 dev->print_log(dev, "%s() Bind error\n", __func__);
405 ishtp_cl_disconnect(cl);
406 return rets;
407 }
408
409 return rets;
410}
411
412/**
413 * ishtp_cl_connect() - Build connection with firmware
414 * @cl: client device instance
415 *
416 * Call ishtp_cl_connect_to_fw() to connect and bind to firmware. If successful,
417 * allocate RX and TX ring buffers, and start flow control with firmware to
418 * start communication.
419 *
420 * Return: 0 if there is successful connection to the firmware, allocate
421 * ring buffers.
422 */
423int ishtp_cl_connect(struct ishtp_cl *cl)
424{
425 struct ishtp_device *dev;
426 int rets;
427
428 if (!cl || !cl->dev)
429 return -ENODEV;
430
431 dev = cl->dev;
432
433 dev->print_log(dev, "%s() current_state = %d\n", __func__, cl->state);
434
435 rets = ishtp_cl_connect_to_fw(cl);
436 if (rets) {
437 dev->print_log(dev, "%s() Connect to fw failed\n", __func__);
438 return rets;
439 }
440
441 rets = ishtp_cl_alloc_rx_ring(cl);
442 if (rets) {
443 dev->print_log(dev, "%s() Alloc RX ring failed\n", __func__);
444 /* if failed allocation, disconnect */
445 ishtp_cl_disconnect(cl);
446 return rets;
447 }
448
449 rets = ishtp_cl_alloc_tx_ring(cl);
450 if (rets) {
451 dev->print_log(dev, "%s() Alloc TX ring failed\n", __func__);
452 /* if failed allocation, disconnect */
453 ishtp_cl_free_rx_ring(cl);
454 ishtp_cl_disconnect(cl);
455 return rets;
456 }
457
458 /*
459 * Upon successful connection and allocation, start flow-control.
460 */
461 rets = ishtp_cl_read_start(cl);
462
463 return rets;
464}
465EXPORT_SYMBOL(ishtp_cl_connect);
466
467/**
468 * ishtp_cl_establish_connection() - Establish connection with the firmware
469 * @cl: client device instance
470 * @uuid: uuid of the client to search
471 * @tx_size: TX ring buffer size
472 * @rx_size: RX ring buffer size
473 * @reset: true if called for reset connection, otherwise for first connection
474 *
475 * This is a helper function for client driver to build connection with firmware.
476 * If it's first time connecting to the firmware, set reset to false, this
477 * function will link client to bus, find client id and send connect request to
478 * the firmware.
479 *
480 * If it's called for reset handler where client lost connection after
481 * firmware reset, set reset to true, this function will reinit client state and
482 * establish connection again. In this case, this function reuses current client
483 * structure and ring buffers to avoid allocation failure and memory fragments.
484 *
485 * Return: 0 for successful connection with the firmware,
486 * or error code on failure
487 */
488int ishtp_cl_establish_connection(struct ishtp_cl *cl, const guid_t *uuid,
489 int tx_size, int rx_size, bool reset)
490{
491 struct ishtp_device *dev;
492 struct ishtp_fw_client *fw_client;
493 int rets;
494
495 if (!cl || !cl->dev)
496 return -ENODEV;
497
498 dev = cl->dev;
499
500 ishtp_set_connection_state(cl, state: ISHTP_CL_INITIALIZING);
501
502 /* reinit ishtp_cl structure if call for reset */
503 if (reset) {
504 cl->host_client_id = 0;
505 cl->fw_client_id = 0;
506 cl->ishtp_flow_ctrl_creds = 0;
507 cl->out_flow_ctrl_creds = 0;
508
509 cl->last_tx_path = CL_TX_PATH_IPC;
510 cl->last_dma_acked = 1;
511 cl->last_dma_addr = NULL;
512 cl->last_ipc_acked = 1;
513
514 cl->sending = 0;
515 cl->err_send_msg = 0;
516 cl->err_send_fc = 0;
517
518 cl->send_msg_cnt_ipc = 0;
519 cl->send_msg_cnt_dma = 0;
520 cl->recv_msg_cnt_ipc = 0;
521 cl->recv_msg_cnt_dma = 0;
522 cl->recv_msg_num_frags = 0;
523 cl->ishtp_flow_ctrl_cnt = 0;
524 cl->out_flow_ctrl_cnt = 0;
525 }
526
527 /* link to bus */
528 rets = ishtp_cl_link(cl);
529 if (rets) {
530 dev->print_log(dev, "%s() ishtp_cl_link failed\n", __func__);
531 return rets;
532 }
533
534 /* find firmware client */
535 fw_client = ishtp_fw_cl_get_client(dev, uuid);
536 if (!fw_client) {
537 dev->print_log(dev,
538 "%s() ish client uuid not found\n", __func__);
539 return -ENOENT;
540 }
541
542 ishtp_set_tx_ring_size(cl, size: tx_size);
543 ishtp_set_rx_ring_size(cl, size: rx_size);
544
545 ishtp_cl_set_fw_client_id(cl, fw_client_id: ishtp_get_fw_client_id(fw_client));
546
547 ishtp_set_connection_state(cl, state: ISHTP_CL_CONNECTING);
548
549 /*
550 * For reset case, not allocate tx/rx ring buffer which are already
551 * done in ishtp_cl_connect() during first connection.
552 */
553 if (reset) {
554 rets = ishtp_cl_connect_to_fw(cl);
555 if (!rets)
556 rets = ishtp_cl_read_start(cl);
557 else
558 dev->print_log(dev,
559 "%s() connect to fw failed\n", __func__);
560 } else {
561 rets = ishtp_cl_connect(cl);
562 }
563
564 return rets;
565}
566EXPORT_SYMBOL(ishtp_cl_establish_connection);
567
568/**
569 * ishtp_cl_destroy_connection() - Disconnect with the firmware
570 * @cl: client device instance
571 * @reset: true if called for firmware reset, false for normal disconnection
572 *
573 * This is a helper function for client driver to disconnect with firmware,
574 * unlink to bus and flush message queue.
575 */
576void ishtp_cl_destroy_connection(struct ishtp_cl *cl, bool reset)
577{
578 if (!cl)
579 return;
580
581 if (reset) {
582 /*
583 * For reset case, connection is already lost during fw reset.
584 * Just set state to DISCONNECTED is enough.
585 */
586 ishtp_set_connection_state(cl, state: ISHTP_CL_DISCONNECTED);
587 } else {
588 if (cl->state != ISHTP_CL_DISCONNECTED) {
589 ishtp_set_connection_state(cl, state: ISHTP_CL_DISCONNECTING);
590 ishtp_cl_disconnect(cl);
591 }
592 }
593
594 ishtp_cl_unlink(cl);
595 ishtp_cl_flush_queues(cl);
596}
597EXPORT_SYMBOL(ishtp_cl_destroy_connection);
598
599/**
600 * ishtp_cl_read_start() - Prepare to read client message
601 * @cl: client device instance
602 *
603 * Get a free buffer from pool of free read buffers and add to read buffer
604 * pool to add contents. Send a flow control request to firmware to be able
605 * send next message.
606 *
607 * Return: 0 if successful or error code on failure
608 */
609int ishtp_cl_read_start(struct ishtp_cl *cl)
610{
611 struct ishtp_device *dev;
612 struct ishtp_cl_rb *rb;
613 int rets;
614 int i;
615 unsigned long flags;
616 unsigned long dev_flags;
617
618 if (WARN_ON(!cl || !cl->dev))
619 return -ENODEV;
620
621 dev = cl->dev;
622
623 if (cl->state != ISHTP_CL_CONNECTED)
624 return -ENODEV;
625
626 if (dev->dev_state != ISHTP_DEV_ENABLED)
627 return -ENODEV;
628
629 i = ishtp_fw_cl_by_id(dev, client_id: cl->fw_client_id);
630 if (i < 0) {
631 dev_err(&cl->device->dev, "no such fw client %d\n",
632 cl->fw_client_id);
633 return -ENODEV;
634 }
635
636 /* The current rb is the head of the free rb list */
637 spin_lock_irqsave(&cl->free_list_spinlock, flags);
638 if (list_empty(head: &cl->free_rb_list.list)) {
639 dev_warn(&cl->device->dev,
640 "[ishtp-ish] Rx buffers pool is empty\n");
641 rets = -ENOMEM;
642 rb = NULL;
643 spin_unlock_irqrestore(lock: &cl->free_list_spinlock, flags);
644 goto out;
645 }
646 rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, list);
647 list_del_init(entry: &rb->list);
648 spin_unlock_irqrestore(lock: &cl->free_list_spinlock, flags);
649
650 rb->cl = cl;
651 rb->buf_idx = 0;
652
653 INIT_LIST_HEAD(list: &rb->list);
654 rets = 0;
655
656 /*
657 * This must be BEFORE sending flow control -
658 * response in ISR may come too fast...
659 */
660 spin_lock_irqsave(&dev->read_list_spinlock, dev_flags);
661 list_add_tail(new: &rb->list, head: &dev->read_list.list);
662 spin_unlock_irqrestore(lock: &dev->read_list_spinlock, flags: dev_flags);
663 if (ishtp_hbm_cl_flow_control_req(dev, cl)) {
664 rets = -ENODEV;
665 goto out;
666 }
667out:
668 /* if ishtp_hbm_cl_flow_control_req failed, return rb to free list */
669 if (rets && rb) {
670 spin_lock_irqsave(&dev->read_list_spinlock, dev_flags);
671 list_del(entry: &rb->list);
672 spin_unlock_irqrestore(lock: &dev->read_list_spinlock, flags: dev_flags);
673
674 spin_lock_irqsave(&cl->free_list_spinlock, flags);
675 list_add_tail(new: &rb->list, head: &cl->free_rb_list.list);
676 spin_unlock_irqrestore(lock: &cl->free_list_spinlock, flags);
677 }
678 return rets;
679}
680
681/**
682 * ishtp_cl_send() - Send a message to firmware
683 * @cl: client device instance
684 * @buf: message buffer
685 * @length: length of message
686 *
687 * If the client is correct state to send message, this function gets a buffer
688 * from tx ring buffers, copy the message data and call to send the message
689 * using ishtp_cl_send_msg()
690 *
691 * Return: 0 if successful or error code on failure
692 */
693int ishtp_cl_send(struct ishtp_cl *cl, uint8_t *buf, size_t length)
694{
695 struct ishtp_device *dev;
696 int id;
697 struct ishtp_cl_tx_ring *cl_msg;
698 int have_msg_to_send = 0;
699 unsigned long tx_flags, tx_free_flags;
700
701 if (WARN_ON(!cl || !cl->dev))
702 return -ENODEV;
703
704 dev = cl->dev;
705
706 if (cl->state != ISHTP_CL_CONNECTED) {
707 ++cl->err_send_msg;
708 return -EPIPE;
709 }
710
711 if (dev->dev_state != ISHTP_DEV_ENABLED) {
712 ++cl->err_send_msg;
713 return -ENODEV;
714 }
715
716 /* Check if we have fw client device */
717 id = ishtp_fw_cl_by_id(dev, client_id: cl->fw_client_id);
718 if (id < 0) {
719 ++cl->err_send_msg;
720 return -ENOENT;
721 }
722
723 if (length > dev->fw_clients[id].props.max_msg_length) {
724 ++cl->err_send_msg;
725 return -EMSGSIZE;
726 }
727
728 /* No free bufs */
729 spin_lock_irqsave(&cl->tx_free_list_spinlock, tx_free_flags);
730 if (list_empty(head: &cl->tx_free_list.list)) {
731 spin_unlock_irqrestore(lock: &cl->tx_free_list_spinlock,
732 flags: tx_free_flags);
733 ++cl->err_send_msg;
734 return -ENOMEM;
735 }
736
737 cl_msg = list_first_entry(&cl->tx_free_list.list,
738 struct ishtp_cl_tx_ring, list);
739 if (!cl_msg->send_buf.data) {
740 spin_unlock_irqrestore(lock: &cl->tx_free_list_spinlock,
741 flags: tx_free_flags);
742 return -EIO;
743 /* Should not happen, as free list is pre-allocated */
744 }
745 /*
746 * This is safe, as 'length' is already checked for not exceeding
747 * max ISHTP message size per client
748 */
749 list_del_init(entry: &cl_msg->list);
750 --cl->tx_ring_free_size;
751
752 spin_unlock_irqrestore(lock: &cl->tx_free_list_spinlock, flags: tx_free_flags);
753 memcpy(cl_msg->send_buf.data, buf, length);
754 cl_msg->send_buf.size = length;
755 spin_lock_irqsave(&cl->tx_list_spinlock, tx_flags);
756 have_msg_to_send = !list_empty(head: &cl->tx_list.list);
757 list_add_tail(new: &cl_msg->list, head: &cl->tx_list.list);
758 spin_unlock_irqrestore(lock: &cl->tx_list_spinlock, flags: tx_flags);
759
760 if (!have_msg_to_send && cl->ishtp_flow_ctrl_creds > 0)
761 ishtp_cl_send_msg(dev, cl);
762
763 return 0;
764}
765EXPORT_SYMBOL(ishtp_cl_send);
766
767/**
768 * ishtp_cl_read_complete() - read complete
769 * @rb: Pointer to client request block
770 *
771 * If the message is completely received call ishtp_cl_bus_rx_event()
772 * to process message
773 */
774static void ishtp_cl_read_complete(struct ishtp_cl_rb *rb)
775{
776 unsigned long flags;
777 int schedule_work_flag = 0;
778 struct ishtp_cl *cl = rb->cl;
779
780 spin_lock_irqsave(&cl->in_process_spinlock, flags);
781 /*
782 * if in-process list is empty, then need to schedule
783 * the processing thread
784 */
785 schedule_work_flag = list_empty(head: &cl->in_process_list.list);
786 list_add_tail(new: &rb->list, head: &cl->in_process_list.list);
787 spin_unlock_irqrestore(lock: &cl->in_process_spinlock, flags);
788
789 if (schedule_work_flag)
790 ishtp_cl_bus_rx_event(device: cl->device);
791}
792
793/**
794 * ipc_tx_send() - IPC tx send function
795 * @prm: Pointer to client device instance
796 *
797 * Send message over IPC. Message will be split into fragments
798 * if message size is bigger than IPC FIFO size, and all
799 * fragments will be sent one by one.
800 */
801static void ipc_tx_send(void *prm)
802{
803 struct ishtp_cl *cl = prm;
804 struct ishtp_cl_tx_ring *cl_msg;
805 size_t rem;
806 struct ishtp_device *dev = (cl ? cl->dev : NULL);
807 struct ishtp_msg_hdr ishtp_hdr;
808 unsigned long tx_flags, tx_free_flags;
809 unsigned char *pmsg;
810
811 if (!dev)
812 return;
813
814 /*
815 * Other conditions if some critical error has
816 * occurred before this callback is called
817 */
818 if (dev->dev_state != ISHTP_DEV_ENABLED)
819 return;
820
821 if (cl->state != ISHTP_CL_CONNECTED)
822 return;
823
824 spin_lock_irqsave(&cl->tx_list_spinlock, tx_flags);
825 if (list_empty(head: &cl->tx_list.list)) {
826 spin_unlock_irqrestore(lock: &cl->tx_list_spinlock, flags: tx_flags);
827 return;
828 }
829
830 if (cl->ishtp_flow_ctrl_creds != 1 && !cl->sending) {
831 spin_unlock_irqrestore(lock: &cl->tx_list_spinlock, flags: tx_flags);
832 return;
833 }
834
835 if (!cl->sending) {
836 --cl->ishtp_flow_ctrl_creds;
837 cl->last_ipc_acked = 0;
838 cl->last_tx_path = CL_TX_PATH_IPC;
839 cl->sending = 1;
840 }
841
842 cl_msg = list_entry(cl->tx_list.list.next, struct ishtp_cl_tx_ring,
843 list);
844 rem = cl_msg->send_buf.size - cl->tx_offs;
845
846 while (rem > 0) {
847 ishtp_hdr.host_addr = cl->host_client_id;
848 ishtp_hdr.fw_addr = cl->fw_client_id;
849 ishtp_hdr.reserved = 0;
850 pmsg = cl_msg->send_buf.data + cl->tx_offs;
851
852 if (rem <= dev->mtu) {
853 /* Last fragment or only one packet */
854 ishtp_hdr.length = rem;
855 ishtp_hdr.msg_complete = 1;
856 /* Submit to IPC queue with no callback */
857 ishtp_write_message(dev, hdr: &ishtp_hdr, buf: pmsg);
858 cl->tx_offs = 0;
859 cl->sending = 0;
860
861 break;
862 } else {
863 /* Send ipc fragment */
864 ishtp_hdr.length = dev->mtu;
865 ishtp_hdr.msg_complete = 0;
866 /* All fregments submitted to IPC queue with no callback */
867 ishtp_write_message(dev, hdr: &ishtp_hdr, buf: pmsg);
868 cl->tx_offs += dev->mtu;
869 rem = cl_msg->send_buf.size - cl->tx_offs;
870 }
871 }
872
873 list_del_init(entry: &cl_msg->list);
874 spin_unlock_irqrestore(lock: &cl->tx_list_spinlock, flags: tx_flags);
875
876 spin_lock_irqsave(&cl->tx_free_list_spinlock, tx_free_flags);
877 list_add_tail(new: &cl_msg->list, head: &cl->tx_free_list.list);
878 ++cl->tx_ring_free_size;
879 spin_unlock_irqrestore(lock: &cl->tx_free_list_spinlock,
880 flags: tx_free_flags);
881}
882
883/**
884 * ishtp_cl_send_msg_ipc() -Send message using IPC
885 * @dev: ISHTP device instance
886 * @cl: Pointer to client device instance
887 *
888 * Send message over IPC not using DMA
889 */
890static void ishtp_cl_send_msg_ipc(struct ishtp_device *dev,
891 struct ishtp_cl *cl)
892{
893 /* If last DMA message wasn't acked yet, leave this one in Tx queue */
894 if (cl->last_tx_path == CL_TX_PATH_DMA && cl->last_dma_acked == 0)
895 return;
896
897 cl->tx_offs = 0;
898 ipc_tx_send(prm: cl);
899 ++cl->send_msg_cnt_ipc;
900}
901
902/**
903 * ishtp_cl_send_msg_dma() -Send message using DMA
904 * @dev: ISHTP device instance
905 * @cl: Pointer to client device instance
906 *
907 * Send message using DMA
908 */
909static void ishtp_cl_send_msg_dma(struct ishtp_device *dev,
910 struct ishtp_cl *cl)
911{
912 struct ishtp_msg_hdr hdr;
913 struct dma_xfer_hbm dma_xfer;
914 unsigned char *msg_addr;
915 int off;
916 struct ishtp_cl_tx_ring *cl_msg;
917 unsigned long tx_flags, tx_free_flags;
918
919 /* If last IPC message wasn't acked yet, leave this one in Tx queue */
920 if (cl->last_tx_path == CL_TX_PATH_IPC && cl->last_ipc_acked == 0)
921 return;
922
923 spin_lock_irqsave(&cl->tx_list_spinlock, tx_flags);
924 if (list_empty(head: &cl->tx_list.list)) {
925 spin_unlock_irqrestore(lock: &cl->tx_list_spinlock, flags: tx_flags);
926 return;
927 }
928
929 cl_msg = list_entry(cl->tx_list.list.next, struct ishtp_cl_tx_ring,
930 list);
931
932 msg_addr = ishtp_cl_get_dma_send_buf(dev, size: cl_msg->send_buf.size);
933 if (!msg_addr) {
934 spin_unlock_irqrestore(lock: &cl->tx_list_spinlock, flags: tx_flags);
935 if (dev->transfer_path == CL_TX_PATH_DEFAULT)
936 ishtp_cl_send_msg_ipc(dev, cl);
937 return;
938 }
939
940 list_del_init(entry: &cl_msg->list); /* Must be before write */
941 spin_unlock_irqrestore(lock: &cl->tx_list_spinlock, flags: tx_flags);
942
943 --cl->ishtp_flow_ctrl_creds;
944 cl->last_dma_acked = 0;
945 cl->last_dma_addr = msg_addr;
946 cl->last_tx_path = CL_TX_PATH_DMA;
947
948 /* write msg to dma buf */
949 memcpy(msg_addr, cl_msg->send_buf.data, cl_msg->send_buf.size);
950
951 /*
952 * if current fw don't support cache snooping, driver have to
953 * flush the cache manually.
954 */
955 if (dev->ops->dma_no_cache_snooping &&
956 dev->ops->dma_no_cache_snooping(dev))
957 clflush_cache_range(addr: msg_addr, size: cl_msg->send_buf.size);
958
959 /* send dma_xfer hbm msg */
960 off = msg_addr - (unsigned char *)dev->ishtp_host_dma_tx_buf;
961 ishtp_hbm_hdr(hdr: &hdr, length: sizeof(struct dma_xfer_hbm));
962 dma_xfer.hbm = DMA_XFER;
963 dma_xfer.fw_client_id = cl->fw_client_id;
964 dma_xfer.host_client_id = cl->host_client_id;
965 dma_xfer.reserved = 0;
966 dma_xfer.msg_addr = dev->ishtp_host_dma_tx_buf_phys + off;
967 dma_xfer.msg_length = cl_msg->send_buf.size;
968 dma_xfer.reserved2 = 0;
969 ishtp_write_message(dev, hdr: &hdr, buf: (unsigned char *)&dma_xfer);
970 spin_lock_irqsave(&cl->tx_free_list_spinlock, tx_free_flags);
971 list_add_tail(new: &cl_msg->list, head: &cl->tx_free_list.list);
972 ++cl->tx_ring_free_size;
973 spin_unlock_irqrestore(lock: &cl->tx_free_list_spinlock, flags: tx_free_flags);
974 ++cl->send_msg_cnt_dma;
975}
976
977/**
978 * ishtp_cl_send_msg() -Send message using DMA or IPC
979 * @dev: ISHTP device instance
980 * @cl: Pointer to client device instance
981 *
982 * Send message using DMA or IPC based on transfer_path
983 */
984void ishtp_cl_send_msg(struct ishtp_device *dev, struct ishtp_cl *cl)
985{
986 if (dev->transfer_path == CL_TX_PATH_DMA)
987 ishtp_cl_send_msg_dma(dev, cl);
988 else
989 ishtp_cl_send_msg_ipc(dev, cl);
990}
991
992/**
993 * recv_ishtp_cl_msg() -Receive client message
994 * @dev: ISHTP device instance
995 * @ishtp_hdr: Pointer to message header
996 *
997 * Receive and dispatch ISHTP client messages. This function executes in ISR
998 * or work queue context
999 */
1000void recv_ishtp_cl_msg(struct ishtp_device *dev,
1001 struct ishtp_msg_hdr *ishtp_hdr)
1002{
1003 struct ishtp_cl *cl;
1004 struct ishtp_cl_rb *rb;
1005 struct ishtp_cl_rb *new_rb;
1006 unsigned char *buffer = NULL;
1007 struct ishtp_cl_rb *complete_rb = NULL;
1008 unsigned long flags;
1009
1010 if (ishtp_hdr->reserved) {
1011 dev_err(dev->devc, "corrupted message header.\n");
1012 goto eoi;
1013 }
1014
1015 if (ishtp_hdr->length > IPC_PAYLOAD_SIZE) {
1016 dev_err(dev->devc,
1017 "ISHTP message length in hdr exceeds IPC MTU\n");
1018 goto eoi;
1019 }
1020
1021 spin_lock_irqsave(&dev->read_list_spinlock, flags);
1022 list_for_each_entry(rb, &dev->read_list.list, list) {
1023 cl = rb->cl;
1024 if (!cl || !(cl->host_client_id == ishtp_hdr->host_addr &&
1025 cl->fw_client_id == ishtp_hdr->fw_addr) ||
1026 !(cl->state == ISHTP_CL_CONNECTED))
1027 continue;
1028
1029 /* If no Rx buffer is allocated, disband the rb */
1030 if (rb->buffer.size == 0 || rb->buffer.data == NULL) {
1031 spin_unlock_irqrestore(lock: &dev->read_list_spinlock, flags);
1032 dev_err(&cl->device->dev,
1033 "Rx buffer is not allocated.\n");
1034 list_del(entry: &rb->list);
1035 ishtp_io_rb_free(priv_rb: rb);
1036 cl->status = -ENOMEM;
1037 goto eoi;
1038 }
1039
1040 /*
1041 * If message buffer overflown (exceeds max. client msg
1042 * size, drop message and return to free buffer.
1043 * Do we need to disconnect such a client? (We don't send
1044 * back FC, so communication will be stuck anyway)
1045 */
1046 if (rb->buffer.size < ishtp_hdr->length + rb->buf_idx) {
1047 spin_unlock_irqrestore(lock: &dev->read_list_spinlock, flags);
1048 dev_err(&cl->device->dev,
1049 "message overflow. size %d len %d idx %ld\n",
1050 rb->buffer.size, ishtp_hdr->length,
1051 rb->buf_idx);
1052 list_del(entry: &rb->list);
1053 ishtp_cl_io_rb_recycle(rb);
1054 cl->status = -EIO;
1055 goto eoi;
1056 }
1057
1058 buffer = rb->buffer.data + rb->buf_idx;
1059 dev->ops->ishtp_read(dev, buffer, ishtp_hdr->length);
1060
1061 rb->buf_idx += ishtp_hdr->length;
1062 if (ishtp_hdr->msg_complete) {
1063 /* Last fragment in message - it's complete */
1064 cl->status = 0;
1065 list_del(entry: &rb->list);
1066 complete_rb = rb;
1067
1068 --cl->out_flow_ctrl_creds;
1069 /*
1070 * the whole msg arrived, send a new FC, and add a new
1071 * rb buffer for the next coming msg
1072 */
1073 spin_lock(lock: &cl->free_list_spinlock);
1074
1075 if (!list_empty(head: &cl->free_rb_list.list)) {
1076 new_rb = list_entry(cl->free_rb_list.list.next,
1077 struct ishtp_cl_rb, list);
1078 list_del_init(entry: &new_rb->list);
1079 spin_unlock(lock: &cl->free_list_spinlock);
1080 new_rb->cl = cl;
1081 new_rb->buf_idx = 0;
1082 INIT_LIST_HEAD(list: &new_rb->list);
1083 list_add_tail(new: &new_rb->list,
1084 head: &dev->read_list.list);
1085
1086 ishtp_hbm_cl_flow_control_req(dev, cl);
1087 } else {
1088 spin_unlock(lock: &cl->free_list_spinlock);
1089 }
1090 }
1091 /* One more fragment in message (even if this was last) */
1092 ++cl->recv_msg_num_frags;
1093
1094 /*
1095 * We can safely break here (and in BH too),
1096 * a single input message can go only to a single request!
1097 */
1098 break;
1099 }
1100
1101 spin_unlock_irqrestore(lock: &dev->read_list_spinlock, flags);
1102 /* If it's nobody's message, just read and discard it */
1103 if (!buffer) {
1104 uint8_t rd_msg_buf[ISHTP_RD_MSG_BUF_SIZE];
1105
1106 dev_err(dev->devc, "Dropped Rx msg - no request\n");
1107 dev->ops->ishtp_read(dev, rd_msg_buf, ishtp_hdr->length);
1108 goto eoi;
1109 }
1110
1111 if (complete_rb) {
1112 cl = complete_rb->cl;
1113 cl->ts_rx = ktime_get();
1114 ++cl->recv_msg_cnt_ipc;
1115 ishtp_cl_read_complete(rb: complete_rb);
1116 }
1117eoi:
1118 return;
1119}
1120
1121/**
1122 * recv_ishtp_cl_msg_dma() -Receive client message
1123 * @dev: ISHTP device instance
1124 * @msg: message pointer
1125 * @hbm: hbm buffer
1126 *
1127 * Receive and dispatch ISHTP client messages using DMA. This function executes
1128 * in ISR or work queue context
1129 */
1130void recv_ishtp_cl_msg_dma(struct ishtp_device *dev, void *msg,
1131 struct dma_xfer_hbm *hbm)
1132{
1133 struct ishtp_cl *cl;
1134 struct ishtp_cl_rb *rb;
1135 struct ishtp_cl_rb *new_rb;
1136 unsigned char *buffer = NULL;
1137 struct ishtp_cl_rb *complete_rb = NULL;
1138 unsigned long flags;
1139
1140 spin_lock_irqsave(&dev->read_list_spinlock, flags);
1141
1142 list_for_each_entry(rb, &dev->read_list.list, list) {
1143 cl = rb->cl;
1144 if (!cl || !(cl->host_client_id == hbm->host_client_id &&
1145 cl->fw_client_id == hbm->fw_client_id) ||
1146 !(cl->state == ISHTP_CL_CONNECTED))
1147 continue;
1148
1149 /*
1150 * If no Rx buffer is allocated, disband the rb
1151 */
1152 if (rb->buffer.size == 0 || rb->buffer.data == NULL) {
1153 spin_unlock_irqrestore(lock: &dev->read_list_spinlock, flags);
1154 dev_err(&cl->device->dev,
1155 "response buffer is not allocated.\n");
1156 list_del(entry: &rb->list);
1157 ishtp_io_rb_free(priv_rb: rb);
1158 cl->status = -ENOMEM;
1159 goto eoi;
1160 }
1161
1162 /*
1163 * If message buffer overflown (exceeds max. client msg
1164 * size, drop message and return to free buffer.
1165 * Do we need to disconnect such a client? (We don't send
1166 * back FC, so communication will be stuck anyway)
1167 */
1168 if (rb->buffer.size < hbm->msg_length) {
1169 spin_unlock_irqrestore(lock: &dev->read_list_spinlock, flags);
1170 dev_err(&cl->device->dev,
1171 "message overflow. size %d len %d idx %ld\n",
1172 rb->buffer.size, hbm->msg_length, rb->buf_idx);
1173 list_del(entry: &rb->list);
1174 ishtp_cl_io_rb_recycle(rb);
1175 cl->status = -EIO;
1176 goto eoi;
1177 }
1178
1179 buffer = rb->buffer.data;
1180
1181 /*
1182 * if current fw don't support cache snooping, driver have to
1183 * flush the cache manually.
1184 */
1185 if (dev->ops->dma_no_cache_snooping &&
1186 dev->ops->dma_no_cache_snooping(dev))
1187 clflush_cache_range(addr: msg, size: hbm->msg_length);
1188
1189 memcpy(buffer, msg, hbm->msg_length);
1190 rb->buf_idx = hbm->msg_length;
1191
1192 /* Last fragment in message - it's complete */
1193 cl->status = 0;
1194 list_del(entry: &rb->list);
1195 complete_rb = rb;
1196
1197 --cl->out_flow_ctrl_creds;
1198 /*
1199 * the whole msg arrived, send a new FC, and add a new
1200 * rb buffer for the next coming msg
1201 */
1202 spin_lock(lock: &cl->free_list_spinlock);
1203
1204 if (!list_empty(head: &cl->free_rb_list.list)) {
1205 new_rb = list_entry(cl->free_rb_list.list.next,
1206 struct ishtp_cl_rb, list);
1207 list_del_init(entry: &new_rb->list);
1208 spin_unlock(lock: &cl->free_list_spinlock);
1209 new_rb->cl = cl;
1210 new_rb->buf_idx = 0;
1211 INIT_LIST_HEAD(list: &new_rb->list);
1212 list_add_tail(new: &new_rb->list,
1213 head: &dev->read_list.list);
1214
1215 ishtp_hbm_cl_flow_control_req(dev, cl);
1216 } else {
1217 spin_unlock(lock: &cl->free_list_spinlock);
1218 }
1219
1220 /* One more fragment in message (this is always last) */
1221 ++cl->recv_msg_num_frags;
1222
1223 /*
1224 * We can safely break here (and in BH too),
1225 * a single input message can go only to a single request!
1226 */
1227 break;
1228 }
1229
1230 spin_unlock_irqrestore(lock: &dev->read_list_spinlock, flags);
1231 /* If it's nobody's message, just read and discard it */
1232 if (!buffer) {
1233 dev_err(dev->devc, "Dropped Rx (DMA) msg - no request\n");
1234 goto eoi;
1235 }
1236
1237 if (complete_rb) {
1238 cl = complete_rb->cl;
1239 cl->ts_rx = ktime_get();
1240 ++cl->recv_msg_cnt_dma;
1241 ishtp_cl_read_complete(rb: complete_rb);
1242 }
1243eoi:
1244 return;
1245}
1246
1247void *ishtp_get_client_data(struct ishtp_cl *cl)
1248{
1249 return cl->client_data;
1250}
1251EXPORT_SYMBOL(ishtp_get_client_data);
1252
1253void ishtp_set_client_data(struct ishtp_cl *cl, void *data)
1254{
1255 cl->client_data = data;
1256}
1257EXPORT_SYMBOL(ishtp_set_client_data);
1258
1259struct ishtp_device *ishtp_get_ishtp_device(struct ishtp_cl *cl)
1260{
1261 return cl->dev;
1262}
1263EXPORT_SYMBOL(ishtp_get_ishtp_device);
1264
1265void ishtp_set_tx_ring_size(struct ishtp_cl *cl, int size)
1266{
1267 cl->tx_ring_size = size;
1268}
1269EXPORT_SYMBOL(ishtp_set_tx_ring_size);
1270
1271void ishtp_set_rx_ring_size(struct ishtp_cl *cl, int size)
1272{
1273 cl->rx_ring_size = size;
1274}
1275EXPORT_SYMBOL(ishtp_set_rx_ring_size);
1276
1277void ishtp_set_connection_state(struct ishtp_cl *cl, int state)
1278{
1279 cl->state = state;
1280}
1281EXPORT_SYMBOL(ishtp_set_connection_state);
1282
1283void ishtp_cl_set_fw_client_id(struct ishtp_cl *cl, int fw_client_id)
1284{
1285 cl->fw_client_id = fw_client_id;
1286}
1287EXPORT_SYMBOL(ishtp_cl_set_fw_client_id);
1288

source code of linux/drivers/hid/intel-ish-hid/ishtp/client.c