1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Copyright (C) 2006 Freescale Semiconductor, Inc. All rights reserved.
4 *
5 * Authors: Shlomi Gridish <gridish@freescale.com>
6 * Li Yang <leoli@freescale.com>
7 *
8 * Description:
9 * QE UCC Fast API Set - UCC Fast specific routines implementations.
10 */
11#include <linux/kernel.h>
12#include <linux/errno.h>
13#include <linux/slab.h>
14#include <linux/stddef.h>
15#include <linux/interrupt.h>
16#include <linux/err.h>
17#include <linux/export.h>
18
19#include <asm/io.h>
20#include <soc/fsl/qe/immap_qe.h>
21#include <soc/fsl/qe/qe.h>
22
23#include <soc/fsl/qe/ucc.h>
24#include <soc/fsl/qe/ucc_fast.h>
25
26void ucc_fast_dump_regs(struct ucc_fast_private * uccf)
27{
28 printk(KERN_INFO "UCC%u Fast registers:\n", uccf->uf_info->ucc_num);
29 printk(KERN_INFO "Base address: 0x%p\n", uccf->uf_regs);
30
31 printk(KERN_INFO "gumr : addr=0x%p, val=0x%08x\n",
32 &uccf->uf_regs->gumr, ioread32be(&uccf->uf_regs->gumr));
33 printk(KERN_INFO "upsmr : addr=0x%p, val=0x%08x\n",
34 &uccf->uf_regs->upsmr, ioread32be(&uccf->uf_regs->upsmr));
35 printk(KERN_INFO "utodr : addr=0x%p, val=0x%04x\n",
36 &uccf->uf_regs->utodr, ioread16be(&uccf->uf_regs->utodr));
37 printk(KERN_INFO "udsr : addr=0x%p, val=0x%04x\n",
38 &uccf->uf_regs->udsr, ioread16be(&uccf->uf_regs->udsr));
39 printk(KERN_INFO "ucce : addr=0x%p, val=0x%08x\n",
40 &uccf->uf_regs->ucce, ioread32be(&uccf->uf_regs->ucce));
41 printk(KERN_INFO "uccm : addr=0x%p, val=0x%08x\n",
42 &uccf->uf_regs->uccm, ioread32be(&uccf->uf_regs->uccm));
43 printk(KERN_INFO "uccs : addr=0x%p, val=0x%02x\n",
44 &uccf->uf_regs->uccs, ioread8(&uccf->uf_regs->uccs));
45 printk(KERN_INFO "urfb : addr=0x%p, val=0x%08x\n",
46 &uccf->uf_regs->urfb, ioread32be(&uccf->uf_regs->urfb));
47 printk(KERN_INFO "urfs : addr=0x%p, val=0x%04x\n",
48 &uccf->uf_regs->urfs, ioread16be(&uccf->uf_regs->urfs));
49 printk(KERN_INFO "urfet : addr=0x%p, val=0x%04x\n",
50 &uccf->uf_regs->urfet, ioread16be(&uccf->uf_regs->urfet));
51 printk(KERN_INFO "urfset: addr=0x%p, val=0x%04x\n",
52 &uccf->uf_regs->urfset,
53 ioread16be(&uccf->uf_regs->urfset));
54 printk(KERN_INFO "utfb : addr=0x%p, val=0x%08x\n",
55 &uccf->uf_regs->utfb, ioread32be(&uccf->uf_regs->utfb));
56 printk(KERN_INFO "utfs : addr=0x%p, val=0x%04x\n",
57 &uccf->uf_regs->utfs, ioread16be(&uccf->uf_regs->utfs));
58 printk(KERN_INFO "utfet : addr=0x%p, val=0x%04x\n",
59 &uccf->uf_regs->utfet, ioread16be(&uccf->uf_regs->utfet));
60 printk(KERN_INFO "utftt : addr=0x%p, val=0x%04x\n",
61 &uccf->uf_regs->utftt, ioread16be(&uccf->uf_regs->utftt));
62 printk(KERN_INFO "utpt : addr=0x%p, val=0x%04x\n",
63 &uccf->uf_regs->utpt, ioread16be(&uccf->uf_regs->utpt));
64 printk(KERN_INFO "urtry : addr=0x%p, val=0x%08x\n",
65 &uccf->uf_regs->urtry, ioread32be(&uccf->uf_regs->urtry));
66 printk(KERN_INFO "guemr : addr=0x%p, val=0x%02x\n",
67 &uccf->uf_regs->guemr, ioread8(&uccf->uf_regs->guemr));
68}
69EXPORT_SYMBOL(ucc_fast_dump_regs);
70
71u32 ucc_fast_get_qe_cr_subblock(int uccf_num)
72{
73 switch (uccf_num) {
74 case 0: return QE_CR_SUBBLOCK_UCCFAST1;
75 case 1: return QE_CR_SUBBLOCK_UCCFAST2;
76 case 2: return QE_CR_SUBBLOCK_UCCFAST3;
77 case 3: return QE_CR_SUBBLOCK_UCCFAST4;
78 case 4: return QE_CR_SUBBLOCK_UCCFAST5;
79 case 5: return QE_CR_SUBBLOCK_UCCFAST6;
80 case 6: return QE_CR_SUBBLOCK_UCCFAST7;
81 case 7: return QE_CR_SUBBLOCK_UCCFAST8;
82 default: return QE_CR_SUBBLOCK_INVALID;
83 }
84}
85EXPORT_SYMBOL(ucc_fast_get_qe_cr_subblock);
86
87void ucc_fast_transmit_on_demand(struct ucc_fast_private * uccf)
88{
89 iowrite16be(UCC_FAST_TOD, &uccf->uf_regs->utodr);
90}
91EXPORT_SYMBOL(ucc_fast_transmit_on_demand);
92
93void ucc_fast_enable(struct ucc_fast_private * uccf, enum comm_dir mode)
94{
95 struct ucc_fast __iomem *uf_regs;
96 u32 gumr;
97
98 uf_regs = uccf->uf_regs;
99
100 /* Enable reception and/or transmission on this UCC. */
101 gumr = ioread32be(&uf_regs->gumr);
102 if (mode & COMM_DIR_TX) {
103 gumr |= UCC_FAST_GUMR_ENT;
104 uccf->enabled_tx = 1;
105 }
106 if (mode & COMM_DIR_RX) {
107 gumr |= UCC_FAST_GUMR_ENR;
108 uccf->enabled_rx = 1;
109 }
110 iowrite32be(gumr, &uf_regs->gumr);
111}
112EXPORT_SYMBOL(ucc_fast_enable);
113
114void ucc_fast_disable(struct ucc_fast_private * uccf, enum comm_dir mode)
115{
116 struct ucc_fast __iomem *uf_regs;
117 u32 gumr;
118
119 uf_regs = uccf->uf_regs;
120
121 /* Disable reception and/or transmission on this UCC. */
122 gumr = ioread32be(&uf_regs->gumr);
123 if (mode & COMM_DIR_TX) {
124 gumr &= ~UCC_FAST_GUMR_ENT;
125 uccf->enabled_tx = 0;
126 }
127 if (mode & COMM_DIR_RX) {
128 gumr &= ~UCC_FAST_GUMR_ENR;
129 uccf->enabled_rx = 0;
130 }
131 iowrite32be(gumr, &uf_regs->gumr);
132}
133EXPORT_SYMBOL(ucc_fast_disable);
134
135int ucc_fast_init(struct ucc_fast_info * uf_info, struct ucc_fast_private ** uccf_ret)
136{
137 struct ucc_fast_private *uccf;
138 struct ucc_fast __iomem *uf_regs;
139 u32 gumr;
140 int ret;
141
142 if (!uf_info)
143 return -EINVAL;
144
145 /* check if the UCC port number is in range. */
146 if ((uf_info->ucc_num < 0) || (uf_info->ucc_num > UCC_MAX_NUM - 1)) {
147 printk(KERN_ERR "%s: illegal UCC number\n", __func__);
148 return -EINVAL;
149 }
150
151 /* Check that 'max_rx_buf_length' is properly aligned (4). */
152 if (uf_info->max_rx_buf_length & (UCC_FAST_MRBLR_ALIGNMENT - 1)) {
153 printk(KERN_ERR "%s: max_rx_buf_length not aligned\n",
154 __func__);
155 return -EINVAL;
156 }
157
158 /* Validate Virtual Fifo register values */
159 if (uf_info->urfs < UCC_FAST_URFS_MIN_VAL) {
160 printk(KERN_ERR "%s: urfs is too small\n", __func__);
161 return -EINVAL;
162 }
163
164 if (uf_info->urfs & (UCC_FAST_VIRT_FIFO_REGS_ALIGNMENT - 1)) {
165 printk(KERN_ERR "%s: urfs is not aligned\n", __func__);
166 return -EINVAL;
167 }
168
169 if (uf_info->urfet & (UCC_FAST_VIRT_FIFO_REGS_ALIGNMENT - 1)) {
170 printk(KERN_ERR "%s: urfet is not aligned.\n", __func__);
171 return -EINVAL;
172 }
173
174 if (uf_info->urfset & (UCC_FAST_VIRT_FIFO_REGS_ALIGNMENT - 1)) {
175 printk(KERN_ERR "%s: urfset is not aligned\n", __func__);
176 return -EINVAL;
177 }
178
179 if (uf_info->utfs & (UCC_FAST_VIRT_FIFO_REGS_ALIGNMENT - 1)) {
180 printk(KERN_ERR "%s: utfs is not aligned\n", __func__);
181 return -EINVAL;
182 }
183
184 if (uf_info->utfet & (UCC_FAST_VIRT_FIFO_REGS_ALIGNMENT - 1)) {
185 printk(KERN_ERR "%s: utfet is not aligned\n", __func__);
186 return -EINVAL;
187 }
188
189 if (uf_info->utftt & (UCC_FAST_VIRT_FIFO_REGS_ALIGNMENT - 1)) {
190 printk(KERN_ERR "%s: utftt is not aligned\n", __func__);
191 return -EINVAL;
192 }
193
194 uccf = kzalloc(size: sizeof(struct ucc_fast_private), GFP_KERNEL);
195 if (!uccf) {
196 printk(KERN_ERR "%s: Cannot allocate private data\n",
197 __func__);
198 return -ENOMEM;
199 }
200 uccf->ucc_fast_tx_virtual_fifo_base_offset = -1;
201 uccf->ucc_fast_rx_virtual_fifo_base_offset = -1;
202
203 /* Fill fast UCC structure */
204 uccf->uf_info = uf_info;
205 /* Set the PHY base address */
206 uccf->uf_regs = ioremap(offset: uf_info->regs, size: sizeof(struct ucc_fast));
207 if (uccf->uf_regs == NULL) {
208 printk(KERN_ERR "%s: Cannot map UCC registers\n", __func__);
209 kfree(objp: uccf);
210 return -ENOMEM;
211 }
212
213 uccf->enabled_tx = 0;
214 uccf->enabled_rx = 0;
215 uccf->stopped_tx = 0;
216 uccf->stopped_rx = 0;
217 uf_regs = uccf->uf_regs;
218 uccf->p_ucce = &uf_regs->ucce;
219 uccf->p_uccm = &uf_regs->uccm;
220#ifdef CONFIG_UGETH_TX_ON_DEMAND
221 uccf->p_utodr = &uf_regs->utodr;
222#endif
223#ifdef STATISTICS
224 uccf->tx_frames = 0;
225 uccf->rx_frames = 0;
226 uccf->rx_discarded = 0;
227#endif /* STATISTICS */
228
229 /* Set UCC to fast type */
230 ret = ucc_set_type(ucc_num: uf_info->ucc_num, speed: UCC_SPEED_TYPE_FAST);
231 if (ret) {
232 printk(KERN_ERR "%s: cannot set UCC type\n", __func__);
233 ucc_fast_free(uccf);
234 return ret;
235 }
236
237 uccf->mrblr = uf_info->max_rx_buf_length;
238
239 /* Set GUMR */
240 /* For more details see the hardware spec. */
241 gumr = uf_info->ttx_trx;
242 if (uf_info->tci)
243 gumr |= UCC_FAST_GUMR_TCI;
244 if (uf_info->cdp)
245 gumr |= UCC_FAST_GUMR_CDP;
246 if (uf_info->ctsp)
247 gumr |= UCC_FAST_GUMR_CTSP;
248 if (uf_info->cds)
249 gumr |= UCC_FAST_GUMR_CDS;
250 if (uf_info->ctss)
251 gumr |= UCC_FAST_GUMR_CTSS;
252 if (uf_info->txsy)
253 gumr |= UCC_FAST_GUMR_TXSY;
254 if (uf_info->rsyn)
255 gumr |= UCC_FAST_GUMR_RSYN;
256 gumr |= uf_info->synl;
257 if (uf_info->rtsm)
258 gumr |= UCC_FAST_GUMR_RTSM;
259 gumr |= uf_info->renc;
260 if (uf_info->revd)
261 gumr |= UCC_FAST_GUMR_REVD;
262 gumr |= uf_info->tenc;
263 gumr |= uf_info->tcrc;
264 gumr |= uf_info->mode;
265 iowrite32be(gumr, &uf_regs->gumr);
266
267 /* Allocate memory for Tx Virtual Fifo */
268 uccf->ucc_fast_tx_virtual_fifo_base_offset =
269 qe_muram_alloc(size: uf_info->utfs, UCC_FAST_VIRT_FIFO_REGS_ALIGNMENT);
270 if (uccf->ucc_fast_tx_virtual_fifo_base_offset < 0) {
271 printk(KERN_ERR "%s: cannot allocate MURAM for TX FIFO\n",
272 __func__);
273 ucc_fast_free(uccf);
274 return -ENOMEM;
275 }
276
277 /* Allocate memory for Rx Virtual Fifo */
278 uccf->ucc_fast_rx_virtual_fifo_base_offset =
279 qe_muram_alloc(size: uf_info->urfs +
280 UCC_FAST_RECEIVE_VIRTUAL_FIFO_SIZE_FUDGE_FACTOR,
281 UCC_FAST_VIRT_FIFO_REGS_ALIGNMENT);
282 if (uccf->ucc_fast_rx_virtual_fifo_base_offset < 0) {
283 printk(KERN_ERR "%s: cannot allocate MURAM for RX FIFO\n",
284 __func__);
285 ucc_fast_free(uccf);
286 return -ENOMEM;
287 }
288
289 /* Set Virtual Fifo registers */
290 iowrite16be(uf_info->urfs, &uf_regs->urfs);
291 iowrite16be(uf_info->urfet, &uf_regs->urfet);
292 iowrite16be(uf_info->urfset, &uf_regs->urfset);
293 iowrite16be(uf_info->utfs, &uf_regs->utfs);
294 iowrite16be(uf_info->utfet, &uf_regs->utfet);
295 iowrite16be(uf_info->utftt, &uf_regs->utftt);
296 /* utfb, urfb are offsets from MURAM base */
297 iowrite32be(uccf->ucc_fast_tx_virtual_fifo_base_offset,
298 &uf_regs->utfb);
299 iowrite32be(uccf->ucc_fast_rx_virtual_fifo_base_offset,
300 &uf_regs->urfb);
301
302 /* Mux clocking */
303 /* Grant Support */
304 ucc_set_qe_mux_grant(ucc_num: uf_info->ucc_num, set: uf_info->grant_support);
305 /* Breakpoint Support */
306 ucc_set_qe_mux_bkpt(ucc_num: uf_info->ucc_num, set: uf_info->brkpt_support);
307 /* Set Tsa or NMSI mode. */
308 ucc_set_qe_mux_tsa(ucc_num: uf_info->ucc_num, set: uf_info->tsa);
309 /* If NMSI (not Tsa), set Tx and Rx clock. */
310 if (!uf_info->tsa) {
311 /* Rx clock routing */
312 if ((uf_info->rx_clock != QE_CLK_NONE) &&
313 ucc_set_qe_mux_rxtx(ucc_num: uf_info->ucc_num, clock: uf_info->rx_clock,
314 mode: COMM_DIR_RX)) {
315 printk(KERN_ERR "%s: illegal value for RX clock\n",
316 __func__);
317 ucc_fast_free(uccf);
318 return -EINVAL;
319 }
320 /* Tx clock routing */
321 if ((uf_info->tx_clock != QE_CLK_NONE) &&
322 ucc_set_qe_mux_rxtx(ucc_num: uf_info->ucc_num, clock: uf_info->tx_clock,
323 mode: COMM_DIR_TX)) {
324 printk(KERN_ERR "%s: illegal value for TX clock\n",
325 __func__);
326 ucc_fast_free(uccf);
327 return -EINVAL;
328 }
329 } else {
330 /* tdm Rx clock routing */
331 if ((uf_info->rx_clock != QE_CLK_NONE) &&
332 ucc_set_tdm_rxtx_clk(tdm_num: uf_info->tdm_num, clock: uf_info->rx_clock,
333 mode: COMM_DIR_RX)) {
334 pr_err("%s: illegal value for RX clock", __func__);
335 ucc_fast_free(uccf);
336 return -EINVAL;
337 }
338
339 /* tdm Tx clock routing */
340 if ((uf_info->tx_clock != QE_CLK_NONE) &&
341 ucc_set_tdm_rxtx_clk(tdm_num: uf_info->tdm_num, clock: uf_info->tx_clock,
342 mode: COMM_DIR_TX)) {
343 pr_err("%s: illegal value for TX clock", __func__);
344 ucc_fast_free(uccf);
345 return -EINVAL;
346 }
347
348 /* tdm Rx sync clock routing */
349 if ((uf_info->rx_sync != QE_CLK_NONE) &&
350 ucc_set_tdm_rxtx_sync(tdm_num: uf_info->tdm_num, clock: uf_info->rx_sync,
351 mode: COMM_DIR_RX)) {
352 pr_err("%s: illegal value for RX clock", __func__);
353 ucc_fast_free(uccf);
354 return -EINVAL;
355 }
356
357 /* tdm Tx sync clock routing */
358 if ((uf_info->tx_sync != QE_CLK_NONE) &&
359 ucc_set_tdm_rxtx_sync(tdm_num: uf_info->tdm_num, clock: uf_info->tx_sync,
360 mode: COMM_DIR_TX)) {
361 pr_err("%s: illegal value for TX clock", __func__);
362 ucc_fast_free(uccf);
363 return -EINVAL;
364 }
365 }
366
367 /* Set interrupt mask register at UCC level. */
368 iowrite32be(uf_info->uccm_mask, &uf_regs->uccm);
369
370 /* First, clear anything pending at UCC level,
371 * otherwise, old garbage may come through
372 * as soon as the dam is opened. */
373
374 /* Writing '1' clears */
375 iowrite32be(0xffffffff, &uf_regs->ucce);
376
377 *uccf_ret = uccf;
378 return 0;
379}
380EXPORT_SYMBOL(ucc_fast_init);
381
382void ucc_fast_free(struct ucc_fast_private * uccf)
383{
384 if (!uccf)
385 return;
386
387 qe_muram_free(offset: uccf->ucc_fast_tx_virtual_fifo_base_offset);
388 qe_muram_free(offset: uccf->ucc_fast_rx_virtual_fifo_base_offset);
389
390 if (uccf->uf_regs)
391 iounmap(addr: uccf->uf_regs);
392
393 kfree(objp: uccf);
394}
395EXPORT_SYMBOL(ucc_fast_free);
396

source code of linux/drivers/soc/fsl/qe/ucc_fast.c