1// SPDX-License-Identifier: GPL-2.0
2/*
3 * SD/MMC Greybus driver.
4 *
5 * Copyright 2014-2015 Google Inc.
6 * Copyright 2014-2015 Linaro Ltd.
7 */
8
9#include <linux/kernel.h>
10#include <linux/mmc/core.h>
11#include <linux/mmc/host.h>
12#include <linux/mmc/mmc.h>
13#include <linux/scatterlist.h>
14#include <linux/workqueue.h>
15#include <linux/greybus.h>
16
17#include "gbphy.h"
18
19struct gb_sdio_host {
20 struct gb_connection *connection;
21 struct gbphy_device *gbphy_dev;
22 struct mmc_host *mmc;
23 struct mmc_request *mrq;
24 struct mutex lock; /* lock for this host */
25 size_t data_max;
26 spinlock_t xfer; /* lock to cancel ongoing transfer */
27 bool xfer_stop;
28 struct workqueue_struct *mrq_workqueue;
29 struct work_struct mrqwork;
30 u8 queued_events;
31 bool removed;
32 bool card_present;
33 bool read_only;
34};
35
36#define GB_SDIO_RSP_R1_R5_R6_R7 (GB_SDIO_RSP_PRESENT | GB_SDIO_RSP_CRC | \
37 GB_SDIO_RSP_OPCODE)
38#define GB_SDIO_RSP_R3_R4 (GB_SDIO_RSP_PRESENT)
39#define GB_SDIO_RSP_R2 (GB_SDIO_RSP_PRESENT | GB_SDIO_RSP_CRC | \
40 GB_SDIO_RSP_136)
41#define GB_SDIO_RSP_R1B (GB_SDIO_RSP_PRESENT | GB_SDIO_RSP_CRC | \
42 GB_SDIO_RSP_OPCODE | GB_SDIO_RSP_BUSY)
43
44/* kernel vdd starts at 0x80 and we need to translate to greybus ones 0x01 */
45#define GB_SDIO_VDD_SHIFT 8
46
47#ifndef MMC_CAP2_CORE_RUNTIME_PM
48#define MMC_CAP2_CORE_RUNTIME_PM 0
49#endif
50
51static inline bool single_op(struct mmc_command *cmd)
52{
53 u32 opcode = cmd->opcode;
54
55 return opcode == MMC_WRITE_BLOCK ||
56 opcode == MMC_READ_SINGLE_BLOCK;
57}
58
59static void _gb_sdio_set_host_caps(struct gb_sdio_host *host, u32 r)
60{
61 u32 caps = 0;
62 u32 caps2 = 0;
63
64 caps = ((r & GB_SDIO_CAP_NONREMOVABLE) ? MMC_CAP_NONREMOVABLE : 0) |
65 ((r & GB_SDIO_CAP_4_BIT_DATA) ? MMC_CAP_4_BIT_DATA : 0) |
66 ((r & GB_SDIO_CAP_8_BIT_DATA) ? MMC_CAP_8_BIT_DATA : 0) |
67 ((r & GB_SDIO_CAP_MMC_HS) ? MMC_CAP_MMC_HIGHSPEED : 0) |
68 ((r & GB_SDIO_CAP_SD_HS) ? MMC_CAP_SD_HIGHSPEED : 0) |
69 ((r & GB_SDIO_CAP_1_2V_DDR) ? MMC_CAP_1_2V_DDR : 0) |
70 ((r & GB_SDIO_CAP_1_8V_DDR) ? MMC_CAP_1_8V_DDR : 0) |
71 ((r & GB_SDIO_CAP_POWER_OFF_CARD) ? MMC_CAP_POWER_OFF_CARD : 0) |
72 ((r & GB_SDIO_CAP_UHS_SDR12) ? MMC_CAP_UHS_SDR12 : 0) |
73 ((r & GB_SDIO_CAP_UHS_SDR25) ? MMC_CAP_UHS_SDR25 : 0) |
74 ((r & GB_SDIO_CAP_UHS_SDR50) ? MMC_CAP_UHS_SDR50 : 0) |
75 ((r & GB_SDIO_CAP_UHS_SDR104) ? MMC_CAP_UHS_SDR104 : 0) |
76 ((r & GB_SDIO_CAP_UHS_DDR50) ? MMC_CAP_UHS_DDR50 : 0) |
77 ((r & GB_SDIO_CAP_DRIVER_TYPE_A) ? MMC_CAP_DRIVER_TYPE_A : 0) |
78 ((r & GB_SDIO_CAP_DRIVER_TYPE_C) ? MMC_CAP_DRIVER_TYPE_C : 0) |
79 ((r & GB_SDIO_CAP_DRIVER_TYPE_D) ? MMC_CAP_DRIVER_TYPE_D : 0);
80
81 caps2 = ((r & GB_SDIO_CAP_HS200_1_2V) ? MMC_CAP2_HS200_1_2V_SDR : 0) |
82 ((r & GB_SDIO_CAP_HS400_1_2V) ? MMC_CAP2_HS400_1_2V : 0) |
83 ((r & GB_SDIO_CAP_HS400_1_8V) ? MMC_CAP2_HS400_1_8V : 0) |
84 ((r & GB_SDIO_CAP_HS200_1_8V) ? MMC_CAP2_HS200_1_8V_SDR : 0);
85
86 host->mmc->caps = caps;
87 host->mmc->caps2 = caps2 | MMC_CAP2_CORE_RUNTIME_PM;
88
89 if (caps & MMC_CAP_NONREMOVABLE)
90 host->card_present = true;
91}
92
93static u32 _gb_sdio_get_host_ocr(u32 ocr)
94{
95 return (((ocr & GB_SDIO_VDD_165_195) ? MMC_VDD_165_195 : 0) |
96 ((ocr & GB_SDIO_VDD_20_21) ? MMC_VDD_20_21 : 0) |
97 ((ocr & GB_SDIO_VDD_21_22) ? MMC_VDD_21_22 : 0) |
98 ((ocr & GB_SDIO_VDD_22_23) ? MMC_VDD_22_23 : 0) |
99 ((ocr & GB_SDIO_VDD_23_24) ? MMC_VDD_23_24 : 0) |
100 ((ocr & GB_SDIO_VDD_24_25) ? MMC_VDD_24_25 : 0) |
101 ((ocr & GB_SDIO_VDD_25_26) ? MMC_VDD_25_26 : 0) |
102 ((ocr & GB_SDIO_VDD_26_27) ? MMC_VDD_26_27 : 0) |
103 ((ocr & GB_SDIO_VDD_27_28) ? MMC_VDD_27_28 : 0) |
104 ((ocr & GB_SDIO_VDD_28_29) ? MMC_VDD_28_29 : 0) |
105 ((ocr & GB_SDIO_VDD_29_30) ? MMC_VDD_29_30 : 0) |
106 ((ocr & GB_SDIO_VDD_30_31) ? MMC_VDD_30_31 : 0) |
107 ((ocr & GB_SDIO_VDD_31_32) ? MMC_VDD_31_32 : 0) |
108 ((ocr & GB_SDIO_VDD_32_33) ? MMC_VDD_32_33 : 0) |
109 ((ocr & GB_SDIO_VDD_33_34) ? MMC_VDD_33_34 : 0) |
110 ((ocr & GB_SDIO_VDD_34_35) ? MMC_VDD_34_35 : 0) |
111 ((ocr & GB_SDIO_VDD_35_36) ? MMC_VDD_35_36 : 0)
112 );
113}
114
115static int gb_sdio_get_caps(struct gb_sdio_host *host)
116{
117 struct gb_sdio_get_caps_response response;
118 struct mmc_host *mmc = host->mmc;
119 u16 data_max;
120 u32 blksz;
121 u32 ocr;
122 u32 r;
123 int ret;
124
125 ret = gb_operation_sync(connection: host->connection, GB_SDIO_TYPE_GET_CAPABILITIES,
126 NULL, request_size: 0, response: &response, response_size: sizeof(response));
127 if (ret < 0)
128 return ret;
129 r = le32_to_cpu(response.caps);
130
131 _gb_sdio_set_host_caps(host, r);
132
133 /* get the max block size that could fit our payload */
134 data_max = gb_operation_get_payload_size_max(connection: host->connection);
135 data_max = min(data_max - sizeof(struct gb_sdio_transfer_request),
136 data_max - sizeof(struct gb_sdio_transfer_response));
137
138 blksz = min_t(u16, le16_to_cpu(response.max_blk_size), data_max);
139 blksz = max_t(u32, 512, blksz);
140
141 mmc->max_blk_size = rounddown_pow_of_two(blksz);
142 mmc->max_blk_count = le16_to_cpu(response.max_blk_count);
143 host->data_max = data_max;
144
145 /* get ocr supported values */
146 ocr = _gb_sdio_get_host_ocr(le32_to_cpu(response.ocr));
147 mmc->ocr_avail = ocr;
148 mmc->ocr_avail_sdio = mmc->ocr_avail;
149 mmc->ocr_avail_sd = mmc->ocr_avail;
150 mmc->ocr_avail_mmc = mmc->ocr_avail;
151
152 /* get frequency range values */
153 mmc->f_min = le32_to_cpu(response.f_min);
154 mmc->f_max = le32_to_cpu(response.f_max);
155
156 return 0;
157}
158
159static void _gb_queue_event(struct gb_sdio_host *host, u8 event)
160{
161 if (event & GB_SDIO_CARD_INSERTED)
162 host->queued_events &= ~GB_SDIO_CARD_REMOVED;
163 else if (event & GB_SDIO_CARD_REMOVED)
164 host->queued_events &= ~GB_SDIO_CARD_INSERTED;
165
166 host->queued_events |= event;
167}
168
169static int _gb_sdio_process_events(struct gb_sdio_host *host, u8 event)
170{
171 u8 state_changed = 0;
172
173 if (event & GB_SDIO_CARD_INSERTED) {
174 if (host->mmc->caps & MMC_CAP_NONREMOVABLE)
175 return 0;
176 if (host->card_present)
177 return 0;
178 host->card_present = true;
179 state_changed = 1;
180 }
181
182 if (event & GB_SDIO_CARD_REMOVED) {
183 if (host->mmc->caps & MMC_CAP_NONREMOVABLE)
184 return 0;
185 if (!(host->card_present))
186 return 0;
187 host->card_present = false;
188 state_changed = 1;
189 }
190
191 if (event & GB_SDIO_WP)
192 host->read_only = true;
193
194 if (state_changed) {
195 dev_info(mmc_dev(host->mmc), "card %s now event\n",
196 (host->card_present ? "inserted" : "removed"));
197 mmc_detect_change(host->mmc, delay: 0);
198 }
199
200 return 0;
201}
202
203static int gb_sdio_request_handler(struct gb_operation *op)
204{
205 struct gb_sdio_host *host = gb_connection_get_data(connection: op->connection);
206 struct gb_message *request;
207 struct gb_sdio_event_request *payload;
208 u8 type = op->type;
209 int ret = 0;
210 u8 event;
211
212 if (type != GB_SDIO_TYPE_EVENT) {
213 dev_err(mmc_dev(host->mmc),
214 "unsupported unsolicited event: %u\n", type);
215 return -EINVAL;
216 }
217
218 request = op->request;
219
220 if (request->payload_size < sizeof(*payload)) {
221 dev_err(mmc_dev(host->mmc), "wrong event size received (%zu < %zu)\n",
222 request->payload_size, sizeof(*payload));
223 return -EINVAL;
224 }
225
226 payload = request->payload;
227 event = payload->event;
228
229 if (host->removed)
230 _gb_queue_event(host, event);
231 else
232 ret = _gb_sdio_process_events(host, event);
233
234 return ret;
235}
236
237static int gb_sdio_set_ios(struct gb_sdio_host *host,
238 struct gb_sdio_set_ios_request *request)
239{
240 int ret;
241
242 ret = gbphy_runtime_get_sync(gbphy_dev: host->gbphy_dev);
243 if (ret)
244 return ret;
245
246 ret = gb_operation_sync(connection: host->connection, GB_SDIO_TYPE_SET_IOS, request,
247 request_size: sizeof(*request), NULL, response_size: 0);
248
249 gbphy_runtime_put_autosuspend(gbphy_dev: host->gbphy_dev);
250
251 return ret;
252}
253
254static int _gb_sdio_send(struct gb_sdio_host *host, struct mmc_data *data,
255 size_t len, u16 nblocks, off_t skip)
256{
257 struct gb_sdio_transfer_request *request;
258 struct gb_sdio_transfer_response *response;
259 struct gb_operation *operation;
260 struct scatterlist *sg = data->sg;
261 unsigned int sg_len = data->sg_len;
262 size_t copied;
263 u16 send_blksz;
264 u16 send_blocks;
265 int ret;
266
267 WARN_ON(len > host->data_max);
268
269 operation = gb_operation_create(connection: host->connection, GB_SDIO_TYPE_TRANSFER,
270 request_size: len + sizeof(*request),
271 response_size: sizeof(*response), GFP_KERNEL);
272 if (!operation)
273 return -ENOMEM;
274
275 request = operation->request->payload;
276 request->data_flags = data->flags >> 8;
277 request->data_blocks = cpu_to_le16(nblocks);
278 request->data_blksz = cpu_to_le16(data->blksz);
279
280 copied = sg_pcopy_to_buffer(sgl: sg, nents: sg_len, buf: &request->data[0], buflen: len, skip);
281
282 if (copied != len) {
283 ret = -EINVAL;
284 goto err_put_operation;
285 }
286
287 ret = gb_operation_request_send_sync(operation);
288 if (ret < 0)
289 goto err_put_operation;
290
291 response = operation->response->payload;
292
293 send_blocks = le16_to_cpu(response->data_blocks);
294 send_blksz = le16_to_cpu(response->data_blksz);
295
296 if (len != send_blksz * send_blocks) {
297 dev_err(mmc_dev(host->mmc), "send: size received: %zu != %d\n",
298 len, send_blksz * send_blocks);
299 ret = -EINVAL;
300 }
301
302err_put_operation:
303 gb_operation_put(operation);
304
305 return ret;
306}
307
308static int _gb_sdio_recv(struct gb_sdio_host *host, struct mmc_data *data,
309 size_t len, u16 nblocks, off_t skip)
310{
311 struct gb_sdio_transfer_request *request;
312 struct gb_sdio_transfer_response *response;
313 struct gb_operation *operation;
314 struct scatterlist *sg = data->sg;
315 unsigned int sg_len = data->sg_len;
316 size_t copied;
317 u16 recv_blksz;
318 u16 recv_blocks;
319 int ret;
320
321 WARN_ON(len > host->data_max);
322
323 operation = gb_operation_create(connection: host->connection, GB_SDIO_TYPE_TRANSFER,
324 request_size: sizeof(*request),
325 response_size: len + sizeof(*response), GFP_KERNEL);
326 if (!operation)
327 return -ENOMEM;
328
329 request = operation->request->payload;
330 request->data_flags = data->flags >> 8;
331 request->data_blocks = cpu_to_le16(nblocks);
332 request->data_blksz = cpu_to_le16(data->blksz);
333
334 ret = gb_operation_request_send_sync(operation);
335 if (ret < 0)
336 goto err_put_operation;
337
338 response = operation->response->payload;
339 recv_blocks = le16_to_cpu(response->data_blocks);
340 recv_blksz = le16_to_cpu(response->data_blksz);
341
342 if (len != recv_blksz * recv_blocks) {
343 dev_err(mmc_dev(host->mmc), "recv: size received: %d != %zu\n",
344 recv_blksz * recv_blocks, len);
345 ret = -EINVAL;
346 goto err_put_operation;
347 }
348
349 copied = sg_pcopy_from_buffer(sgl: sg, nents: sg_len, buf: &response->data[0], buflen: len,
350 skip);
351 if (copied != len)
352 ret = -EINVAL;
353
354err_put_operation:
355 gb_operation_put(operation);
356
357 return ret;
358}
359
360static int gb_sdio_transfer(struct gb_sdio_host *host, struct mmc_data *data)
361{
362 size_t left, len;
363 off_t skip = 0;
364 int ret = 0;
365 u16 nblocks;
366
367 if (single_op(cmd: data->mrq->cmd) && data->blocks > 1) {
368 ret = -ETIMEDOUT;
369 goto out;
370 }
371
372 left = data->blksz * data->blocks;
373
374 while (left) {
375 /* check is a stop transmission is pending */
376 spin_lock(lock: &host->xfer);
377 if (host->xfer_stop) {
378 host->xfer_stop = false;
379 spin_unlock(lock: &host->xfer);
380 ret = -EINTR;
381 goto out;
382 }
383 spin_unlock(lock: &host->xfer);
384 len = min(left, host->data_max);
385 nblocks = len / data->blksz;
386 len = nblocks * data->blksz;
387
388 if (data->flags & MMC_DATA_READ) {
389 ret = _gb_sdio_recv(host, data, len, nblocks, skip);
390 if (ret < 0)
391 goto out;
392 } else {
393 ret = _gb_sdio_send(host, data, len, nblocks, skip);
394 if (ret < 0)
395 goto out;
396 }
397 data->bytes_xfered += len;
398 left -= len;
399 skip += len;
400 }
401
402out:
403 data->error = ret;
404 return ret;
405}
406
407static int gb_sdio_command(struct gb_sdio_host *host, struct mmc_command *cmd)
408{
409 struct gb_sdio_command_request request = {0};
410 struct gb_sdio_command_response response;
411 struct mmc_data *data = host->mrq->data;
412 unsigned int timeout_ms;
413 u8 cmd_flags;
414 u8 cmd_type;
415 int i;
416 int ret;
417
418 switch (mmc_resp_type(cmd)) {
419 case MMC_RSP_NONE:
420 cmd_flags = GB_SDIO_RSP_NONE;
421 break;
422 case MMC_RSP_R1:
423 cmd_flags = GB_SDIO_RSP_R1_R5_R6_R7;
424 break;
425 case MMC_RSP_R1B:
426 cmd_flags = GB_SDIO_RSP_R1B;
427 break;
428 case MMC_RSP_R2:
429 cmd_flags = GB_SDIO_RSP_R2;
430 break;
431 case MMC_RSP_R3:
432 cmd_flags = GB_SDIO_RSP_R3_R4;
433 break;
434 default:
435 dev_err(mmc_dev(host->mmc), "cmd flag invalid 0x%04x\n",
436 mmc_resp_type(cmd));
437 ret = -EINVAL;
438 goto out;
439 }
440
441 switch (mmc_cmd_type(cmd)) {
442 case MMC_CMD_BC:
443 cmd_type = GB_SDIO_CMD_BC;
444 break;
445 case MMC_CMD_BCR:
446 cmd_type = GB_SDIO_CMD_BCR;
447 break;
448 case MMC_CMD_AC:
449 cmd_type = GB_SDIO_CMD_AC;
450 break;
451 case MMC_CMD_ADTC:
452 cmd_type = GB_SDIO_CMD_ADTC;
453 break;
454 default:
455 dev_err(mmc_dev(host->mmc), "cmd type invalid 0x%04x\n",
456 mmc_cmd_type(cmd));
457 ret = -EINVAL;
458 goto out;
459 }
460
461 request.cmd = cmd->opcode;
462 request.cmd_flags = cmd_flags;
463 request.cmd_type = cmd_type;
464 request.cmd_arg = cpu_to_le32(cmd->arg);
465 /* some controllers need to know at command time data details */
466 if (data) {
467 request.data_blocks = cpu_to_le16(data->blocks);
468 request.data_blksz = cpu_to_le16(data->blksz);
469 }
470
471 timeout_ms = cmd->busy_timeout ? cmd->busy_timeout :
472 GB_OPERATION_TIMEOUT_DEFAULT;
473
474 ret = gb_operation_sync_timeout(connection: host->connection, GB_SDIO_TYPE_COMMAND,
475 request: &request, request_size: sizeof(request), response: &response,
476 response_size: sizeof(response), timeout: timeout_ms);
477 if (ret < 0)
478 goto out;
479
480 /* no response expected */
481 if (cmd_flags == GB_SDIO_RSP_NONE)
482 goto out;
483
484 /* long response expected */
485 if (cmd_flags & GB_SDIO_RSP_R2)
486 for (i = 0; i < 4; i++)
487 cmd->resp[i] = le32_to_cpu(response.resp[i]);
488 else
489 cmd->resp[0] = le32_to_cpu(response.resp[0]);
490
491out:
492 cmd->error = ret;
493 return ret;
494}
495
496static void gb_sdio_mrq_work(struct work_struct *work)
497{
498 struct gb_sdio_host *host;
499 struct mmc_request *mrq;
500 int ret;
501
502 host = container_of(work, struct gb_sdio_host, mrqwork);
503
504 ret = gbphy_runtime_get_sync(gbphy_dev: host->gbphy_dev);
505 if (ret)
506 return;
507
508 mutex_lock(&host->lock);
509 mrq = host->mrq;
510 if (!mrq) {
511 mutex_unlock(lock: &host->lock);
512 gbphy_runtime_put_autosuspend(gbphy_dev: host->gbphy_dev);
513 dev_err(mmc_dev(host->mmc), "mmc request is NULL");
514 return;
515 }
516
517 if (host->removed) {
518 mrq->cmd->error = -ESHUTDOWN;
519 goto done;
520 }
521
522 if (mrq->sbc) {
523 ret = gb_sdio_command(host, cmd: mrq->sbc);
524 if (ret < 0)
525 goto done;
526 }
527
528 ret = gb_sdio_command(host, cmd: mrq->cmd);
529 if (ret < 0)
530 goto done;
531
532 if (mrq->data) {
533 ret = gb_sdio_transfer(host, data: mrq->data);
534 if (ret < 0)
535 goto done;
536 }
537
538 if (mrq->stop) {
539 ret = gb_sdio_command(host, cmd: mrq->stop);
540 if (ret < 0)
541 goto done;
542 }
543
544done:
545 host->mrq = NULL;
546 mutex_unlock(lock: &host->lock);
547 mmc_request_done(host->mmc, mrq);
548 gbphy_runtime_put_autosuspend(gbphy_dev: host->gbphy_dev);
549}
550
551static void gb_mmc_request(struct mmc_host *mmc, struct mmc_request *mrq)
552{
553 struct gb_sdio_host *host = mmc_priv(host: mmc);
554 struct mmc_command *cmd = mrq->cmd;
555
556 /* Check if it is a cancel to ongoing transfer */
557 if (cmd->opcode == MMC_STOP_TRANSMISSION) {
558 spin_lock(lock: &host->xfer);
559 host->xfer_stop = true;
560 spin_unlock(lock: &host->xfer);
561 }
562
563 mutex_lock(&host->lock);
564
565 WARN_ON(host->mrq);
566 host->mrq = mrq;
567
568 if (host->removed) {
569 mrq->cmd->error = -ESHUTDOWN;
570 goto out;
571 }
572 if (!host->card_present) {
573 mrq->cmd->error = -ENOMEDIUM;
574 goto out;
575 }
576
577 queue_work(wq: host->mrq_workqueue, work: &host->mrqwork);
578
579 mutex_unlock(lock: &host->lock);
580 return;
581
582out:
583 host->mrq = NULL;
584 mutex_unlock(lock: &host->lock);
585 mmc_request_done(mmc, mrq);
586}
587
588static void gb_mmc_set_ios(struct mmc_host *mmc, struct mmc_ios *ios)
589{
590 struct gb_sdio_host *host = mmc_priv(host: mmc);
591 struct gb_sdio_set_ios_request request;
592 int ret;
593 u8 power_mode;
594 u8 bus_width;
595 u8 timing;
596 u8 signal_voltage;
597 u8 drv_type;
598 u32 vdd = 0;
599
600 mutex_lock(&host->lock);
601 request.clock = cpu_to_le32(ios->clock);
602
603 if (ios->vdd)
604 vdd = 1 << (ios->vdd - GB_SDIO_VDD_SHIFT);
605 request.vdd = cpu_to_le32(vdd);
606
607 request.bus_mode = ios->bus_mode == MMC_BUSMODE_OPENDRAIN ?
608 GB_SDIO_BUSMODE_OPENDRAIN :
609 GB_SDIO_BUSMODE_PUSHPULL;
610
611 switch (ios->power_mode) {
612 case MMC_POWER_OFF:
613 default:
614 power_mode = GB_SDIO_POWER_OFF;
615 break;
616 case MMC_POWER_UP:
617 power_mode = GB_SDIO_POWER_UP;
618 break;
619 case MMC_POWER_ON:
620 power_mode = GB_SDIO_POWER_ON;
621 break;
622 case MMC_POWER_UNDEFINED:
623 power_mode = GB_SDIO_POWER_UNDEFINED;
624 break;
625 }
626 request.power_mode = power_mode;
627
628 switch (ios->bus_width) {
629 case MMC_BUS_WIDTH_1:
630 bus_width = GB_SDIO_BUS_WIDTH_1;
631 break;
632 case MMC_BUS_WIDTH_4:
633 default:
634 bus_width = GB_SDIO_BUS_WIDTH_4;
635 break;
636 case MMC_BUS_WIDTH_8:
637 bus_width = GB_SDIO_BUS_WIDTH_8;
638 break;
639 }
640 request.bus_width = bus_width;
641
642 switch (ios->timing) {
643 case MMC_TIMING_LEGACY:
644 default:
645 timing = GB_SDIO_TIMING_LEGACY;
646 break;
647 case MMC_TIMING_MMC_HS:
648 timing = GB_SDIO_TIMING_MMC_HS;
649 break;
650 case MMC_TIMING_SD_HS:
651 timing = GB_SDIO_TIMING_SD_HS;
652 break;
653 case MMC_TIMING_UHS_SDR12:
654 timing = GB_SDIO_TIMING_UHS_SDR12;
655 break;
656 case MMC_TIMING_UHS_SDR25:
657 timing = GB_SDIO_TIMING_UHS_SDR25;
658 break;
659 case MMC_TIMING_UHS_SDR50:
660 timing = GB_SDIO_TIMING_UHS_SDR50;
661 break;
662 case MMC_TIMING_UHS_SDR104:
663 timing = GB_SDIO_TIMING_UHS_SDR104;
664 break;
665 case MMC_TIMING_UHS_DDR50:
666 timing = GB_SDIO_TIMING_UHS_DDR50;
667 break;
668 case MMC_TIMING_MMC_DDR52:
669 timing = GB_SDIO_TIMING_MMC_DDR52;
670 break;
671 case MMC_TIMING_MMC_HS200:
672 timing = GB_SDIO_TIMING_MMC_HS200;
673 break;
674 case MMC_TIMING_MMC_HS400:
675 timing = GB_SDIO_TIMING_MMC_HS400;
676 break;
677 }
678 request.timing = timing;
679
680 switch (ios->signal_voltage) {
681 case MMC_SIGNAL_VOLTAGE_330:
682 signal_voltage = GB_SDIO_SIGNAL_VOLTAGE_330;
683 break;
684 case MMC_SIGNAL_VOLTAGE_180:
685 default:
686 signal_voltage = GB_SDIO_SIGNAL_VOLTAGE_180;
687 break;
688 case MMC_SIGNAL_VOLTAGE_120:
689 signal_voltage = GB_SDIO_SIGNAL_VOLTAGE_120;
690 break;
691 }
692 request.signal_voltage = signal_voltage;
693
694 switch (ios->drv_type) {
695 case MMC_SET_DRIVER_TYPE_A:
696 drv_type = GB_SDIO_SET_DRIVER_TYPE_A;
697 break;
698 case MMC_SET_DRIVER_TYPE_C:
699 drv_type = GB_SDIO_SET_DRIVER_TYPE_C;
700 break;
701 case MMC_SET_DRIVER_TYPE_D:
702 drv_type = GB_SDIO_SET_DRIVER_TYPE_D;
703 break;
704 case MMC_SET_DRIVER_TYPE_B:
705 default:
706 drv_type = GB_SDIO_SET_DRIVER_TYPE_B;
707 break;
708 }
709 request.drv_type = drv_type;
710
711 ret = gb_sdio_set_ios(host, request: &request);
712 if (ret < 0)
713 goto out;
714
715 memcpy(&mmc->ios, ios, sizeof(mmc->ios));
716
717out:
718 mutex_unlock(lock: &host->lock);
719}
720
721static int gb_mmc_get_ro(struct mmc_host *mmc)
722{
723 struct gb_sdio_host *host = mmc_priv(host: mmc);
724
725 mutex_lock(&host->lock);
726 if (host->removed) {
727 mutex_unlock(lock: &host->lock);
728 return -ESHUTDOWN;
729 }
730 mutex_unlock(lock: &host->lock);
731
732 return host->read_only;
733}
734
735static int gb_mmc_get_cd(struct mmc_host *mmc)
736{
737 struct gb_sdio_host *host = mmc_priv(host: mmc);
738
739 mutex_lock(&host->lock);
740 if (host->removed) {
741 mutex_unlock(lock: &host->lock);
742 return -ESHUTDOWN;
743 }
744 mutex_unlock(lock: &host->lock);
745
746 return host->card_present;
747}
748
749static int gb_mmc_switch_voltage(struct mmc_host *mmc, struct mmc_ios *ios)
750{
751 return 0;
752}
753
754static const struct mmc_host_ops gb_sdio_ops = {
755 .request = gb_mmc_request,
756 .set_ios = gb_mmc_set_ios,
757 .get_ro = gb_mmc_get_ro,
758 .get_cd = gb_mmc_get_cd,
759 .start_signal_voltage_switch = gb_mmc_switch_voltage,
760};
761
762static int gb_sdio_probe(struct gbphy_device *gbphy_dev,
763 const struct gbphy_device_id *id)
764{
765 struct gb_connection *connection;
766 struct mmc_host *mmc;
767 struct gb_sdio_host *host;
768 int ret = 0;
769
770 mmc = mmc_alloc_host(extra: sizeof(*host), &gbphy_dev->dev);
771 if (!mmc)
772 return -ENOMEM;
773
774 connection = gb_connection_create(bundle: gbphy_dev->bundle,
775 le16_to_cpu(gbphy_dev->cport_desc->id),
776 handler: gb_sdio_request_handler);
777 if (IS_ERR(ptr: connection)) {
778 ret = PTR_ERR(ptr: connection);
779 goto exit_mmc_free;
780 }
781
782 host = mmc_priv(host: mmc);
783 host->mmc = mmc;
784 host->removed = true;
785
786 host->connection = connection;
787 gb_connection_set_data(connection, data: host);
788 host->gbphy_dev = gbphy_dev;
789 gb_gbphy_set_data(gdev: gbphy_dev, data: host);
790
791 ret = gb_connection_enable_tx(connection);
792 if (ret)
793 goto exit_connection_destroy;
794
795 ret = gb_sdio_get_caps(host);
796 if (ret < 0)
797 goto exit_connection_disable;
798
799 mmc->ops = &gb_sdio_ops;
800
801 mmc->max_segs = host->mmc->max_blk_count;
802
803 /* for now we make a map 1:1 between max request and segment size */
804 mmc->max_req_size = mmc->max_blk_size * mmc->max_blk_count;
805 mmc->max_seg_size = mmc->max_req_size;
806
807 mutex_init(&host->lock);
808 spin_lock_init(&host->xfer);
809 host->mrq_workqueue = alloc_workqueue(fmt: "mmc-%s", flags: 0, max_active: 1,
810 dev_name(dev: &gbphy_dev->dev));
811 if (!host->mrq_workqueue) {
812 ret = -ENOMEM;
813 goto exit_connection_disable;
814 }
815 INIT_WORK(&host->mrqwork, gb_sdio_mrq_work);
816
817 ret = gb_connection_enable(connection);
818 if (ret)
819 goto exit_wq_destroy;
820
821 ret = mmc_add_host(mmc);
822 if (ret < 0)
823 goto exit_wq_destroy;
824 host->removed = false;
825 ret = _gb_sdio_process_events(host, event: host->queued_events);
826 host->queued_events = 0;
827
828 gbphy_runtime_put_autosuspend(gbphy_dev);
829
830 return ret;
831
832exit_wq_destroy:
833 destroy_workqueue(wq: host->mrq_workqueue);
834exit_connection_disable:
835 gb_connection_disable(connection);
836exit_connection_destroy:
837 gb_connection_destroy(connection);
838exit_mmc_free:
839 mmc_free_host(mmc);
840
841 return ret;
842}
843
844static void gb_sdio_remove(struct gbphy_device *gbphy_dev)
845{
846 struct gb_sdio_host *host = gb_gbphy_get_data(gdev: gbphy_dev);
847 struct gb_connection *connection = host->connection;
848 struct mmc_host *mmc;
849 int ret;
850
851 ret = gbphy_runtime_get_sync(gbphy_dev);
852 if (ret)
853 gbphy_runtime_get_noresume(gbphy_dev);
854
855 mutex_lock(&host->lock);
856 host->removed = true;
857 mmc = host->mmc;
858 gb_connection_set_data(connection, NULL);
859 mutex_unlock(lock: &host->lock);
860
861 destroy_workqueue(wq: host->mrq_workqueue);
862 gb_connection_disable_rx(connection);
863 mmc_remove_host(mmc);
864 gb_connection_disable(connection);
865 gb_connection_destroy(connection);
866 mmc_free_host(mmc);
867}
868
869static const struct gbphy_device_id gb_sdio_id_table[] = {
870 { GBPHY_PROTOCOL(GREYBUS_PROTOCOL_SDIO) },
871 { },
872};
873MODULE_DEVICE_TABLE(gbphy, gb_sdio_id_table);
874
875static struct gbphy_driver sdio_driver = {
876 .name = "sdio",
877 .probe = gb_sdio_probe,
878 .remove = gb_sdio_remove,
879 .id_table = gb_sdio_id_table,
880};
881
882module_gbphy_driver(sdio_driver);
883MODULE_LICENSE("GPL v2");
884

source code of linux/drivers/staging/greybus/sdio.c