1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Copyright (C) 2020-2022 MaxLinear, Inc.
4 * Copyright (C) 2020 Intel Corporation.
5 * Zhu Yixin <yzhu@maxlinear.com>
6 * Rahul Tanwar <rtanwar@maxlinear.com>
7 */
8#include <linux/clk-provider.h>
9#include <linux/device.h>
10#include <linux/of.h>
11
12#include "clk-cgu.h"
13
14#define GATE_HW_REG_STAT(reg) ((reg) + 0x0)
15#define GATE_HW_REG_EN(reg) ((reg) + 0x4)
16#define GATE_HW_REG_DIS(reg) ((reg) + 0x8)
17#define MAX_DDIV_REG 8
18#define MAX_DIVIDER_VAL 64
19
20#define to_lgm_clk_mux(_hw) container_of(_hw, struct lgm_clk_mux, hw)
21#define to_lgm_clk_divider(_hw) container_of(_hw, struct lgm_clk_divider, hw)
22#define to_lgm_clk_gate(_hw) container_of(_hw, struct lgm_clk_gate, hw)
23#define to_lgm_clk_ddiv(_hw) container_of(_hw, struct lgm_clk_ddiv, hw)
24
25static struct clk_hw *lgm_clk_register_fixed(struct lgm_clk_provider *ctx,
26 const struct lgm_clk_branch *list)
27{
28
29 if (list->div_flags & CLOCK_FLAG_VAL_INIT)
30 lgm_set_clk_val(membase: ctx->membase, reg: list->div_off, shift: list->div_shift,
31 width: list->div_width, set_val: list->div_val);
32
33 return clk_hw_register_fixed_rate(NULL, list->name,
34 list->parent_data[0].name,
35 list->flags, list->mux_flags);
36}
37
38static u8 lgm_clk_mux_get_parent(struct clk_hw *hw)
39{
40 struct lgm_clk_mux *mux = to_lgm_clk_mux(hw);
41 u32 val;
42
43 if (mux->flags & MUX_CLK_SW)
44 val = mux->reg;
45 else
46 val = lgm_get_clk_val(membase: mux->membase, reg: mux->reg, shift: mux->shift,
47 width: mux->width);
48 return clk_mux_val_to_index(hw, NULL, flags: mux->flags, val);
49}
50
51static int lgm_clk_mux_set_parent(struct clk_hw *hw, u8 index)
52{
53 struct lgm_clk_mux *mux = to_lgm_clk_mux(hw);
54 u32 val;
55
56 val = clk_mux_index_to_val(NULL, flags: mux->flags, index);
57 if (mux->flags & MUX_CLK_SW)
58 mux->reg = val;
59 else
60 lgm_set_clk_val(membase: mux->membase, reg: mux->reg, shift: mux->shift,
61 width: mux->width, set_val: val);
62
63 return 0;
64}
65
66static int lgm_clk_mux_determine_rate(struct clk_hw *hw,
67 struct clk_rate_request *req)
68{
69 struct lgm_clk_mux *mux = to_lgm_clk_mux(hw);
70
71 return clk_mux_determine_rate_flags(hw, req, flags: mux->flags);
72}
73
74static const struct clk_ops lgm_clk_mux_ops = {
75 .get_parent = lgm_clk_mux_get_parent,
76 .set_parent = lgm_clk_mux_set_parent,
77 .determine_rate = lgm_clk_mux_determine_rate,
78};
79
80static struct clk_hw *
81lgm_clk_register_mux(struct lgm_clk_provider *ctx,
82 const struct lgm_clk_branch *list)
83{
84 unsigned long cflags = list->mux_flags;
85 struct device *dev = ctx->dev;
86 u8 shift = list->mux_shift;
87 u8 width = list->mux_width;
88 struct clk_init_data init = {};
89 struct lgm_clk_mux *mux;
90 u32 reg = list->mux_off;
91 struct clk_hw *hw;
92 int ret;
93
94 mux = devm_kzalloc(dev, size: sizeof(*mux), GFP_KERNEL);
95 if (!mux)
96 return ERR_PTR(error: -ENOMEM);
97
98 init.name = list->name;
99 init.ops = &lgm_clk_mux_ops;
100 init.flags = list->flags;
101 init.parent_data = list->parent_data;
102 init.num_parents = list->num_parents;
103
104 mux->membase = ctx->membase;
105 mux->reg = reg;
106 mux->shift = shift;
107 mux->width = width;
108 mux->flags = cflags;
109 mux->hw.init = &init;
110
111 hw = &mux->hw;
112 ret = devm_clk_hw_register(dev, hw);
113 if (ret)
114 return ERR_PTR(error: ret);
115
116 if (cflags & CLOCK_FLAG_VAL_INIT)
117 lgm_set_clk_val(membase: mux->membase, reg, shift, width, set_val: list->mux_val);
118
119 return hw;
120}
121
122static unsigned long
123lgm_clk_divider_recalc_rate(struct clk_hw *hw, unsigned long parent_rate)
124{
125 struct lgm_clk_divider *divider = to_lgm_clk_divider(hw);
126 unsigned int val;
127
128 val = lgm_get_clk_val(membase: divider->membase, reg: divider->reg,
129 shift: divider->shift, width: divider->width);
130
131 return divider_recalc_rate(hw, parent_rate, val, table: divider->table,
132 flags: divider->flags, width: divider->width);
133}
134
135static long
136lgm_clk_divider_round_rate(struct clk_hw *hw, unsigned long rate,
137 unsigned long *prate)
138{
139 struct lgm_clk_divider *divider = to_lgm_clk_divider(hw);
140
141 return divider_round_rate(hw, rate, prate, table: divider->table,
142 width: divider->width, flags: divider->flags);
143}
144
145static int
146lgm_clk_divider_set_rate(struct clk_hw *hw, unsigned long rate,
147 unsigned long prate)
148{
149 struct lgm_clk_divider *divider = to_lgm_clk_divider(hw);
150 int value;
151
152 value = divider_get_val(rate, parent_rate: prate, table: divider->table,
153 width: divider->width, flags: divider->flags);
154 if (value < 0)
155 return value;
156
157 lgm_set_clk_val(membase: divider->membase, reg: divider->reg,
158 shift: divider->shift, width: divider->width, set_val: value);
159
160 return 0;
161}
162
163static int lgm_clk_divider_enable_disable(struct clk_hw *hw, int enable)
164{
165 struct lgm_clk_divider *div = to_lgm_clk_divider(hw);
166
167 if (div->flags != DIV_CLK_NO_MASK)
168 lgm_set_clk_val(membase: div->membase, reg: div->reg, shift: div->shift_gate,
169 width: div->width_gate, set_val: enable);
170 return 0;
171}
172
173static int lgm_clk_divider_enable(struct clk_hw *hw)
174{
175 return lgm_clk_divider_enable_disable(hw, enable: 1);
176}
177
178static void lgm_clk_divider_disable(struct clk_hw *hw)
179{
180 lgm_clk_divider_enable_disable(hw, enable: 0);
181}
182
183static const struct clk_ops lgm_clk_divider_ops = {
184 .recalc_rate = lgm_clk_divider_recalc_rate,
185 .round_rate = lgm_clk_divider_round_rate,
186 .set_rate = lgm_clk_divider_set_rate,
187 .enable = lgm_clk_divider_enable,
188 .disable = lgm_clk_divider_disable,
189};
190
191static struct clk_hw *
192lgm_clk_register_divider(struct lgm_clk_provider *ctx,
193 const struct lgm_clk_branch *list)
194{
195 unsigned long cflags = list->div_flags;
196 struct device *dev = ctx->dev;
197 struct lgm_clk_divider *div;
198 struct clk_init_data init = {};
199 u8 shift = list->div_shift;
200 u8 width = list->div_width;
201 u8 shift_gate = list->div_shift_gate;
202 u8 width_gate = list->div_width_gate;
203 u32 reg = list->div_off;
204 struct clk_hw *hw;
205 int ret;
206
207 div = devm_kzalloc(dev, size: sizeof(*div), GFP_KERNEL);
208 if (!div)
209 return ERR_PTR(error: -ENOMEM);
210
211 init.name = list->name;
212 init.ops = &lgm_clk_divider_ops;
213 init.flags = list->flags;
214 init.parent_data = list->parent_data;
215 init.num_parents = 1;
216
217 div->membase = ctx->membase;
218 div->reg = reg;
219 div->shift = shift;
220 div->width = width;
221 div->shift_gate = shift_gate;
222 div->width_gate = width_gate;
223 div->flags = cflags;
224 div->table = list->div_table;
225 div->hw.init = &init;
226
227 hw = &div->hw;
228 ret = devm_clk_hw_register(dev, hw);
229 if (ret)
230 return ERR_PTR(error: ret);
231
232 if (cflags & CLOCK_FLAG_VAL_INIT)
233 lgm_set_clk_val(membase: div->membase, reg, shift, width, set_val: list->div_val);
234
235 return hw;
236}
237
238static struct clk_hw *
239lgm_clk_register_fixed_factor(struct lgm_clk_provider *ctx,
240 const struct lgm_clk_branch *list)
241{
242 struct clk_hw *hw;
243
244 hw = clk_hw_register_fixed_factor(dev: ctx->dev, name: list->name,
245 parent_name: list->parent_data[0].name, flags: list->flags,
246 mult: list->mult, div: list->div);
247 if (IS_ERR(ptr: hw))
248 return ERR_CAST(ptr: hw);
249
250 if (list->div_flags & CLOCK_FLAG_VAL_INIT)
251 lgm_set_clk_val(membase: ctx->membase, reg: list->div_off, shift: list->div_shift,
252 width: list->div_width, set_val: list->div_val);
253
254 return hw;
255}
256
257static int lgm_clk_gate_enable(struct clk_hw *hw)
258{
259 struct lgm_clk_gate *gate = to_lgm_clk_gate(hw);
260 unsigned int reg;
261
262 reg = GATE_HW_REG_EN(gate->reg);
263 lgm_set_clk_val(membase: gate->membase, reg, shift: gate->shift, width: 1, set_val: 1);
264
265 return 0;
266}
267
268static void lgm_clk_gate_disable(struct clk_hw *hw)
269{
270 struct lgm_clk_gate *gate = to_lgm_clk_gate(hw);
271 unsigned int reg;
272
273 reg = GATE_HW_REG_DIS(gate->reg);
274 lgm_set_clk_val(membase: gate->membase, reg, shift: gate->shift, width: 1, set_val: 1);
275}
276
277static int lgm_clk_gate_is_enabled(struct clk_hw *hw)
278{
279 struct lgm_clk_gate *gate = to_lgm_clk_gate(hw);
280 unsigned int reg, ret;
281
282 reg = GATE_HW_REG_STAT(gate->reg);
283 ret = lgm_get_clk_val(membase: gate->membase, reg, shift: gate->shift, width: 1);
284
285 return ret;
286}
287
288static const struct clk_ops lgm_clk_gate_ops = {
289 .enable = lgm_clk_gate_enable,
290 .disable = lgm_clk_gate_disable,
291 .is_enabled = lgm_clk_gate_is_enabled,
292};
293
294static struct clk_hw *
295lgm_clk_register_gate(struct lgm_clk_provider *ctx,
296 const struct lgm_clk_branch *list)
297{
298 unsigned long cflags = list->gate_flags;
299 const char *pname = list->parent_data[0].name;
300 struct device *dev = ctx->dev;
301 u8 shift = list->gate_shift;
302 struct clk_init_data init = {};
303 struct lgm_clk_gate *gate;
304 u32 reg = list->gate_off;
305 struct clk_hw *hw;
306 int ret;
307
308 gate = devm_kzalloc(dev, size: sizeof(*gate), GFP_KERNEL);
309 if (!gate)
310 return ERR_PTR(error: -ENOMEM);
311
312 init.name = list->name;
313 init.ops = &lgm_clk_gate_ops;
314 init.flags = list->flags;
315 init.parent_names = pname ? &pname : NULL;
316 init.num_parents = pname ? 1 : 0;
317
318 gate->membase = ctx->membase;
319 gate->reg = reg;
320 gate->shift = shift;
321 gate->flags = cflags;
322 gate->hw.init = &init;
323
324 hw = &gate->hw;
325 ret = devm_clk_hw_register(dev, hw);
326 if (ret)
327 return ERR_PTR(error: ret);
328
329 if (cflags & CLOCK_FLAG_VAL_INIT) {
330 lgm_set_clk_val(membase: gate->membase, reg, shift, width: 1, set_val: list->gate_val);
331 }
332
333 return hw;
334}
335
336int lgm_clk_register_branches(struct lgm_clk_provider *ctx,
337 const struct lgm_clk_branch *list,
338 unsigned int nr_clk)
339{
340 struct clk_hw *hw;
341 unsigned int idx;
342
343 for (idx = 0; idx < nr_clk; idx++, list++) {
344 switch (list->type) {
345 case CLK_TYPE_FIXED:
346 hw = lgm_clk_register_fixed(ctx, list);
347 break;
348 case CLK_TYPE_MUX:
349 hw = lgm_clk_register_mux(ctx, list);
350 break;
351 case CLK_TYPE_DIVIDER:
352 hw = lgm_clk_register_divider(ctx, list);
353 break;
354 case CLK_TYPE_FIXED_FACTOR:
355 hw = lgm_clk_register_fixed_factor(ctx, list);
356 break;
357 case CLK_TYPE_GATE:
358 if (list->gate_flags & GATE_CLK_HW) {
359 hw = lgm_clk_register_gate(ctx, list);
360 } else {
361 /*
362 * GATE_CLKs can be controlled either from
363 * CGU clk driver i.e. this driver or directly
364 * from power management driver/daemon. It is
365 * dependent on the power policy/profile requirements
366 * of the end product. To override control of gate
367 * clks from this driver, provide NULL for this index
368 * of gate clk provider.
369 */
370 hw = NULL;
371 }
372 break;
373
374 default:
375 dev_err(ctx->dev, "invalid clk type\n");
376 return -EINVAL;
377 }
378
379 if (IS_ERR(ptr: hw)) {
380 dev_err(ctx->dev,
381 "register clk: %s, type: %u failed!\n",
382 list->name, list->type);
383 return -EIO;
384 }
385 ctx->clk_data.hws[list->id] = hw;
386 }
387
388 return 0;
389}
390
391static unsigned long
392lgm_clk_ddiv_recalc_rate(struct clk_hw *hw, unsigned long parent_rate)
393{
394 struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
395 unsigned int div0, div1, exdiv;
396 u64 prate;
397
398 div0 = lgm_get_clk_val(membase: ddiv->membase, reg: ddiv->reg,
399 shift: ddiv->shift0, width: ddiv->width0) + 1;
400 div1 = lgm_get_clk_val(membase: ddiv->membase, reg: ddiv->reg,
401 shift: ddiv->shift1, width: ddiv->width1) + 1;
402 exdiv = lgm_get_clk_val(membase: ddiv->membase, reg: ddiv->reg,
403 shift: ddiv->shift2, width: ddiv->width2);
404 prate = (u64)parent_rate;
405 do_div(prate, div0);
406 do_div(prate, div1);
407
408 if (exdiv) {
409 do_div(prate, ddiv->div);
410 prate *= ddiv->mult;
411 }
412
413 return prate;
414}
415
416static int lgm_clk_ddiv_enable(struct clk_hw *hw)
417{
418 struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
419
420 lgm_set_clk_val(membase: ddiv->membase, reg: ddiv->reg, shift: ddiv->shift_gate,
421 width: ddiv->width_gate, set_val: 1);
422 return 0;
423}
424
425static void lgm_clk_ddiv_disable(struct clk_hw *hw)
426{
427 struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
428
429 lgm_set_clk_val(membase: ddiv->membase, reg: ddiv->reg, shift: ddiv->shift_gate,
430 width: ddiv->width_gate, set_val: 0);
431}
432
433static int
434lgm_clk_get_ddiv_val(u32 div, u32 *ddiv1, u32 *ddiv2)
435{
436 u32 idx, temp;
437
438 *ddiv1 = 1;
439 *ddiv2 = 1;
440
441 if (div > MAX_DIVIDER_VAL)
442 div = MAX_DIVIDER_VAL;
443
444 if (div > 1) {
445 for (idx = 2; idx <= MAX_DDIV_REG; idx++) {
446 temp = DIV_ROUND_UP_ULL((u64)div, idx);
447 if (div % idx == 0 && temp <= MAX_DDIV_REG)
448 break;
449 }
450
451 if (idx > MAX_DDIV_REG)
452 return -EINVAL;
453
454 *ddiv1 = temp;
455 *ddiv2 = idx;
456 }
457
458 return 0;
459}
460
461static int
462lgm_clk_ddiv_set_rate(struct clk_hw *hw, unsigned long rate,
463 unsigned long prate)
464{
465 struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
466 u32 div, ddiv1, ddiv2;
467
468 div = DIV_ROUND_CLOSEST_ULL((u64)prate, rate);
469
470 if (lgm_get_clk_val(membase: ddiv->membase, reg: ddiv->reg, shift: ddiv->shift2, width: 1)) {
471 div = DIV_ROUND_CLOSEST_ULL((u64)div, 5);
472 div = div * 2;
473 }
474
475 if (div <= 0)
476 return -EINVAL;
477
478 if (lgm_clk_get_ddiv_val(div, ddiv1: &ddiv1, ddiv2: &ddiv2))
479 return -EINVAL;
480
481 lgm_set_clk_val(membase: ddiv->membase, reg: ddiv->reg, shift: ddiv->shift0, width: ddiv->width0,
482 set_val: ddiv1 - 1);
483
484 lgm_set_clk_val(membase: ddiv->membase, reg: ddiv->reg, shift: ddiv->shift1, width: ddiv->width1,
485 set_val: ddiv2 - 1);
486
487 return 0;
488}
489
490static long
491lgm_clk_ddiv_round_rate(struct clk_hw *hw, unsigned long rate,
492 unsigned long *prate)
493{
494 struct lgm_clk_ddiv *ddiv = to_lgm_clk_ddiv(hw);
495 u32 div, ddiv1, ddiv2;
496 u64 rate64;
497
498 div = DIV_ROUND_CLOSEST_ULL((u64)*prate, rate);
499
500 /* if predivide bit is enabled, modify div by factor of 2.5 */
501 if (lgm_get_clk_val(membase: ddiv->membase, reg: ddiv->reg, shift: ddiv->shift2, width: 1)) {
502 div = div * 2;
503 div = DIV_ROUND_CLOSEST_ULL((u64)div, 5);
504 }
505
506 if (div <= 0)
507 return *prate;
508
509 if (lgm_clk_get_ddiv_val(div, ddiv1: &ddiv1, ddiv2: &ddiv2) != 0)
510 if (lgm_clk_get_ddiv_val(div: div + 1, ddiv1: &ddiv1, ddiv2: &ddiv2) != 0)
511 return -EINVAL;
512
513 rate64 = *prate;
514 do_div(rate64, ddiv1);
515 do_div(rate64, ddiv2);
516
517 /* if predivide bit is enabled, modify rounded rate by factor of 2.5 */
518 if (lgm_get_clk_val(membase: ddiv->membase, reg: ddiv->reg, shift: ddiv->shift2, width: 1)) {
519 rate64 = rate64 * 2;
520 rate64 = DIV_ROUND_CLOSEST_ULL(rate64, 5);
521 }
522
523 return rate64;
524}
525
526static const struct clk_ops lgm_clk_ddiv_ops = {
527 .recalc_rate = lgm_clk_ddiv_recalc_rate,
528 .enable = lgm_clk_ddiv_enable,
529 .disable = lgm_clk_ddiv_disable,
530 .set_rate = lgm_clk_ddiv_set_rate,
531 .round_rate = lgm_clk_ddiv_round_rate,
532};
533
534int lgm_clk_register_ddiv(struct lgm_clk_provider *ctx,
535 const struct lgm_clk_ddiv_data *list,
536 unsigned int nr_clk)
537{
538 struct device *dev = ctx->dev;
539 struct clk_hw *hw;
540 unsigned int idx;
541 int ret;
542
543 for (idx = 0; idx < nr_clk; idx++, list++) {
544 struct clk_init_data init = {};
545 struct lgm_clk_ddiv *ddiv;
546
547 ddiv = devm_kzalloc(dev, size: sizeof(*ddiv), GFP_KERNEL);
548 if (!ddiv)
549 return -ENOMEM;
550
551 init.name = list->name;
552 init.ops = &lgm_clk_ddiv_ops;
553 init.flags = list->flags;
554 init.parent_data = list->parent_data;
555 init.num_parents = 1;
556
557 ddiv->membase = ctx->membase;
558 ddiv->reg = list->reg;
559 ddiv->shift0 = list->shift0;
560 ddiv->width0 = list->width0;
561 ddiv->shift1 = list->shift1;
562 ddiv->width1 = list->width1;
563 ddiv->shift_gate = list->shift_gate;
564 ddiv->width_gate = list->width_gate;
565 ddiv->shift2 = list->ex_shift;
566 ddiv->width2 = list->ex_width;
567 ddiv->flags = list->div_flags;
568 ddiv->mult = 2;
569 ddiv->div = 5;
570 ddiv->hw.init = &init;
571
572 hw = &ddiv->hw;
573 ret = devm_clk_hw_register(dev, hw);
574 if (ret) {
575 dev_err(dev, "register clk: %s failed!\n", list->name);
576 return ret;
577 }
578 ctx->clk_data.hws[list->id] = hw;
579 }
580
581 return 0;
582}
583

source code of linux/drivers/clk/x86/clk-cgu.c