blob: 482f0937cb5a94947e919dedcdca68a29638cac2 [file] [log] [blame]
Tom Rini83d290c2018-05-06 17:58:06 -04001// SPDX-License-Identifier: GPL-2.0+
Stefan Herbrechtsmeier3a64b252017-01-17 16:27:29 +01002/*
3 * Copyright (C) 2017 Weidmüller Interface GmbH & Co. KG
4 * Stefan Herbrechtsmeier <stefan.herbrechtsmeier@weidmueller.com>
5 *
6 * Copyright (C) 2013 Soren Brinkmann <soren.brinkmann@xilinx.com>
7 * Copyright (C) 2013 Xilinx, Inc. All rights reserved.
Stefan Herbrechtsmeier3a64b252017-01-17 16:27:29 +01008 */
9
10#include <common.h>
11#include <clk-uclass.h>
12#include <dm.h>
13#include <dm/lists.h>
14#include <errno.h>
15#include <asm/io.h>
16#include <asm/arch/clk.h>
17#include <asm/arch/hardware.h>
18#include <asm/arch/sys_proto.h>
19
20/* Register bitfield defines */
21#define PLLCTRL_FBDIV_MASK 0x7f000
22#define PLLCTRL_FBDIV_SHIFT 12
23#define PLLCTRL_BPFORCE_MASK (1 << 4)
24#define PLLCTRL_PWRDWN_MASK 2
25#define PLLCTRL_PWRDWN_SHIFT 1
26#define PLLCTRL_RESET_MASK 1
27#define PLLCTRL_RESET_SHIFT 0
28
29#define ZYNQ_CLK_MAXDIV 0x3f
30#define CLK_CTRL_DIV1_SHIFT 20
31#define CLK_CTRL_DIV1_MASK (ZYNQ_CLK_MAXDIV << CLK_CTRL_DIV1_SHIFT)
32#define CLK_CTRL_DIV0_SHIFT 8
33#define CLK_CTRL_DIV0_MASK (ZYNQ_CLK_MAXDIV << CLK_CTRL_DIV0_SHIFT)
34#define CLK_CTRL_SRCSEL_SHIFT 4
35#define CLK_CTRL_SRCSEL_MASK (0x3 << CLK_CTRL_SRCSEL_SHIFT)
36
37#define CLK_CTRL_DIV2X_SHIFT 26
38#define CLK_CTRL_DIV2X_MASK (ZYNQ_CLK_MAXDIV << CLK_CTRL_DIV2X_SHIFT)
39#define CLK_CTRL_DIV3X_SHIFT 20
40#define CLK_CTRL_DIV3X_MASK (ZYNQ_CLK_MAXDIV << CLK_CTRL_DIV3X_SHIFT)
41
42DECLARE_GLOBAL_DATA_PTR;
43
44#ifndef CONFIG_SPL_BUILD
45enum zynq_clk_rclk {mio_clk, emio_clk};
46#endif
47
48struct zynq_clk_priv {
49 ulong ps_clk_freq;
Stefan Herbrechtsmeier9bb803d2017-01-17 16:27:31 +010050#ifndef CONFIG_SPL_BUILD
51 struct clk gem_emio_clk[2];
52#endif
Stefan Herbrechtsmeier3a64b252017-01-17 16:27:29 +010053};
54
55static void *zynq_clk_get_register(enum zynq_clk id)
56{
57 switch (id) {
58 case armpll_clk:
59 return &slcr_base->arm_pll_ctrl;
60 case ddrpll_clk:
61 return &slcr_base->ddr_pll_ctrl;
62 case iopll_clk:
63 return &slcr_base->io_pll_ctrl;
64 case lqspi_clk:
65 return &slcr_base->lqspi_clk_ctrl;
66 case smc_clk:
67 return &slcr_base->smc_clk_ctrl;
68 case pcap_clk:
69 return &slcr_base->pcap_clk_ctrl;
70 case sdio0_clk ... sdio1_clk:
71 return &slcr_base->sdio_clk_ctrl;
72 case uart0_clk ... uart1_clk:
73 return &slcr_base->uart_clk_ctrl;
74 case spi0_clk ... spi1_clk:
75 return &slcr_base->spi_clk_ctrl;
76#ifndef CONFIG_SPL_BUILD
77 case dci_clk:
78 return &slcr_base->dci_clk_ctrl;
79 case gem0_clk:
80 return &slcr_base->gem0_clk_ctrl;
81 case gem1_clk:
82 return &slcr_base->gem1_clk_ctrl;
83 case fclk0_clk:
84 return &slcr_base->fpga0_clk_ctrl;
85 case fclk1_clk:
86 return &slcr_base->fpga1_clk_ctrl;
87 case fclk2_clk:
88 return &slcr_base->fpga2_clk_ctrl;
89 case fclk3_clk:
90 return &slcr_base->fpga3_clk_ctrl;
91 case can0_clk ... can1_clk:
92 return &slcr_base->can_clk_ctrl;
93 case dbg_trc_clk ... dbg_apb_clk:
94 /* fall through */
95#endif
96 default:
97 return &slcr_base->dbg_clk_ctrl;
98 }
99}
100
101static enum zynq_clk zynq_clk_get_cpu_pll(u32 clk_ctrl)
102{
103 u32 srcsel = (clk_ctrl & CLK_CTRL_SRCSEL_MASK) >> CLK_CTRL_SRCSEL_SHIFT;
104
105 switch (srcsel) {
106 case 2:
107 return ddrpll_clk;
108 case 3:
109 return iopll_clk;
110 case 0 ... 1:
111 default:
112 return armpll_clk;
113 }
114}
115
116static enum zynq_clk zynq_clk_get_peripheral_pll(u32 clk_ctrl)
117{
118 u32 srcsel = (clk_ctrl & CLK_CTRL_SRCSEL_MASK) >> CLK_CTRL_SRCSEL_SHIFT;
119
120 switch (srcsel) {
121 case 2:
122 return armpll_clk;
123 case 3:
124 return ddrpll_clk;
125 case 0 ... 1:
126 default:
127 return iopll_clk;
128 }
129}
130
131static ulong zynq_clk_get_pll_rate(struct zynq_clk_priv *priv, enum zynq_clk id)
132{
133 u32 clk_ctrl, reset, pwrdwn, mul, bypass;
134
135 clk_ctrl = readl(zynq_clk_get_register(id));
136
137 reset = (clk_ctrl & PLLCTRL_RESET_MASK) >> PLLCTRL_RESET_SHIFT;
138 pwrdwn = (clk_ctrl & PLLCTRL_PWRDWN_MASK) >> PLLCTRL_PWRDWN_SHIFT;
139 if (reset || pwrdwn)
140 return 0;
141
142 bypass = clk_ctrl & PLLCTRL_BPFORCE_MASK;
143 if (bypass)
144 mul = 1;
145 else
146 mul = (clk_ctrl & PLLCTRL_FBDIV_MASK) >> PLLCTRL_FBDIV_SHIFT;
147
148 return priv->ps_clk_freq * mul;
149}
150
151#ifndef CONFIG_SPL_BUILD
152static enum zynq_clk_rclk zynq_clk_get_gem_rclk(enum zynq_clk id)
153{
154 u32 clk_ctrl, srcsel;
155
156 if (id == gem0_clk)
157 clk_ctrl = readl(&slcr_base->gem0_rclk_ctrl);
158 else
159 clk_ctrl = readl(&slcr_base->gem1_rclk_ctrl);
160
161 srcsel = (clk_ctrl & CLK_CTRL_SRCSEL_MASK) >> CLK_CTRL_SRCSEL_SHIFT;
162 if (srcsel)
163 return emio_clk;
164 else
165 return mio_clk;
166}
167#endif
168
169static ulong zynq_clk_get_cpu_rate(struct zynq_clk_priv *priv, enum zynq_clk id)
170{
171 u32 clk_621, clk_ctrl, div;
172 enum zynq_clk pll;
173
174 clk_ctrl = readl(&slcr_base->arm_clk_ctrl);
175
176 div = (clk_ctrl & CLK_CTRL_DIV0_MASK) >> CLK_CTRL_DIV0_SHIFT;
177
178 switch (id) {
179 case cpu_1x_clk:
180 div *= 2;
181 /* fall through */
182 case cpu_2x_clk:
183 clk_621 = readl(&slcr_base->clk_621_true) & 1;
184 div *= 2 + clk_621;
185 break;
186 case cpu_3or2x_clk:
187 div *= 2;
188 /* fall through */
189 case cpu_6or4x_clk:
190 break;
191 default:
192 return 0;
193 }
194
195 pll = zynq_clk_get_cpu_pll(clk_ctrl);
196
197 return DIV_ROUND_CLOSEST(zynq_clk_get_pll_rate(priv, pll), div);
198}
199
200#ifndef CONFIG_SPL_BUILD
201static ulong zynq_clk_get_ddr2x_rate(struct zynq_clk_priv *priv)
202{
203 u32 clk_ctrl, div;
204
205 clk_ctrl = readl(&slcr_base->ddr_clk_ctrl);
206
207 div = (clk_ctrl & CLK_CTRL_DIV2X_MASK) >> CLK_CTRL_DIV2X_SHIFT;
208
209 return DIV_ROUND_CLOSEST(zynq_clk_get_pll_rate(priv, ddrpll_clk), div);
210}
211#endif
212
213static ulong zynq_clk_get_ddr3x_rate(struct zynq_clk_priv *priv)
214{
215 u32 clk_ctrl, div;
216
217 clk_ctrl = readl(&slcr_base->ddr_clk_ctrl);
218
219 div = (clk_ctrl & CLK_CTRL_DIV3X_MASK) >> CLK_CTRL_DIV3X_SHIFT;
220
221 return DIV_ROUND_CLOSEST(zynq_clk_get_pll_rate(priv, ddrpll_clk), div);
222}
223
224#ifndef CONFIG_SPL_BUILD
225static ulong zynq_clk_get_dci_rate(struct zynq_clk_priv *priv)
226{
227 u32 clk_ctrl, div0, div1;
228
229 clk_ctrl = readl(&slcr_base->dci_clk_ctrl);
230
231 div0 = (clk_ctrl & CLK_CTRL_DIV0_MASK) >> CLK_CTRL_DIV0_SHIFT;
232 div1 = (clk_ctrl & CLK_CTRL_DIV1_MASK) >> CLK_CTRL_DIV1_SHIFT;
233
234 return DIV_ROUND_CLOSEST(DIV_ROUND_CLOSEST(
235 zynq_clk_get_pll_rate(priv, ddrpll_clk), div0), div1);
236}
237#endif
238
239static ulong zynq_clk_get_peripheral_rate(struct zynq_clk_priv *priv,
240 enum zynq_clk id, bool two_divs)
241{
242 enum zynq_clk pll;
243 u32 clk_ctrl, div0;
244 u32 div1 = 1;
245
246 clk_ctrl = readl(zynq_clk_get_register(id));
247
248 div0 = (clk_ctrl & CLK_CTRL_DIV0_MASK) >> CLK_CTRL_DIV0_SHIFT;
249 if (!div0)
250 div0 = 1;
251
252#ifndef CONFIG_SPL_BUILD
253 if (two_divs) {
254 div1 = (clk_ctrl & CLK_CTRL_DIV1_MASK) >> CLK_CTRL_DIV1_SHIFT;
255 if (!div1)
256 div1 = 1;
257 }
258#endif
259
260 pll = zynq_clk_get_peripheral_pll(clk_ctrl);
261
262 return
263 DIV_ROUND_CLOSEST(
264 DIV_ROUND_CLOSEST(
265 zynq_clk_get_pll_rate(priv, pll), div0),
266 div1);
267}
268
269#ifndef CONFIG_SPL_BUILD
270static ulong zynq_clk_get_gem_rate(struct zynq_clk_priv *priv, enum zynq_clk id)
271{
Stefan Herbrechtsmeier9bb803d2017-01-17 16:27:31 +0100272 struct clk *parent;
273
Stefan Herbrechtsmeier3a64b252017-01-17 16:27:29 +0100274 if (zynq_clk_get_gem_rclk(id) == mio_clk)
275 return zynq_clk_get_peripheral_rate(priv, id, true);
276
Stefan Herbrechtsmeier9bb803d2017-01-17 16:27:31 +0100277 parent = &priv->gem_emio_clk[id - gem0_clk];
278 if (parent->dev)
279 return clk_get_rate(parent);
280
Stefan Herbrechtsmeier3a64b252017-01-17 16:27:29 +0100281 debug("%s: gem%d emio rx clock source unknown\n", __func__,
282 id - gem0_clk);
283
284 return -ENOSYS;
285}
286
287static unsigned long zynq_clk_calc_peripheral_two_divs(ulong rate,
288 ulong pll_rate,
289 u32 *div0, u32 *div1)
290{
291 long new_err, best_err = (long)(~0UL >> 1);
292 ulong new_rate, best_rate = 0;
293 u32 d0, d1;
294
295 for (d0 = 1; d0 <= ZYNQ_CLK_MAXDIV; d0++) {
296 for (d1 = 1; d1 <= ZYNQ_CLK_MAXDIV >> 1; d1++) {
297 new_rate = DIV_ROUND_CLOSEST(
298 DIV_ROUND_CLOSEST(pll_rate, d0), d1);
299 new_err = abs(new_rate - rate);
300
301 if (new_err < best_err) {
302 *div0 = d0;
303 *div1 = d1;
304 best_err = new_err;
305 best_rate = new_rate;
306 }
307 }
308 }
309
310 return best_rate;
311}
312
313static ulong zynq_clk_set_peripheral_rate(struct zynq_clk_priv *priv,
314 enum zynq_clk id, ulong rate,
315 bool two_divs)
316{
317 enum zynq_clk pll;
318 u32 clk_ctrl, div0 = 0, div1 = 0;
319 ulong pll_rate, new_rate;
320 u32 *reg;
321
322 reg = zynq_clk_get_register(id);
323 clk_ctrl = readl(reg);
324
325 pll = zynq_clk_get_peripheral_pll(clk_ctrl);
326 pll_rate = zynq_clk_get_pll_rate(priv, pll);
327 clk_ctrl &= ~CLK_CTRL_DIV0_MASK;
328 if (two_divs) {
329 clk_ctrl &= ~CLK_CTRL_DIV1_MASK;
330 new_rate = zynq_clk_calc_peripheral_two_divs(rate, pll_rate,
331 &div0, &div1);
332 clk_ctrl |= div1 << CLK_CTRL_DIV1_SHIFT;
333 } else {
334 div0 = DIV_ROUND_CLOSEST(pll_rate, rate);
335 if (div0 > ZYNQ_CLK_MAXDIV)
336 div0 = ZYNQ_CLK_MAXDIV;
337 new_rate = DIV_ROUND_CLOSEST(rate, div0);
338 }
339 clk_ctrl |= div0 << CLK_CTRL_DIV0_SHIFT;
340
341 zynq_slcr_unlock();
342 writel(clk_ctrl, reg);
343 zynq_slcr_lock();
344
345 return new_rate;
346}
347
348static ulong zynq_clk_set_gem_rate(struct zynq_clk_priv *priv, enum zynq_clk id,
349 ulong rate)
350{
Stefan Herbrechtsmeier9bb803d2017-01-17 16:27:31 +0100351 struct clk *parent;
352
Stefan Herbrechtsmeier3a64b252017-01-17 16:27:29 +0100353 if (zynq_clk_get_gem_rclk(id) == mio_clk)
354 return zynq_clk_set_peripheral_rate(priv, id, rate, true);
355
Stefan Herbrechtsmeier9bb803d2017-01-17 16:27:31 +0100356 parent = &priv->gem_emio_clk[id - gem0_clk];
357 if (parent->dev)
358 return clk_set_rate(parent, rate);
359
Stefan Herbrechtsmeier3a64b252017-01-17 16:27:29 +0100360 debug("%s: gem%d emio rx clock source unknown\n", __func__,
361 id - gem0_clk);
362
363 return -ENOSYS;
364}
365#endif
366
367#ifndef CONFIG_SPL_BUILD
368static ulong zynq_clk_get_rate(struct clk *clk)
369{
370 struct zynq_clk_priv *priv = dev_get_priv(clk->dev);
371 enum zynq_clk id = clk->id;
372 bool two_divs = false;
373
374 switch (id) {
375 case armpll_clk ... iopll_clk:
376 return zynq_clk_get_pll_rate(priv, id);
377 case cpu_6or4x_clk ... cpu_1x_clk:
378 return zynq_clk_get_cpu_rate(priv, id);
379 case ddr2x_clk:
380 return zynq_clk_get_ddr2x_rate(priv);
381 case ddr3x_clk:
382 return zynq_clk_get_ddr3x_rate(priv);
383 case dci_clk:
384 return zynq_clk_get_dci_rate(priv);
385 case gem0_clk ... gem1_clk:
386 return zynq_clk_get_gem_rate(priv, id);
387 case fclk0_clk ... can1_clk:
388 two_divs = true;
389 /* fall through */
390 case dbg_trc_clk ... dbg_apb_clk:
391 case lqspi_clk ... pcap_clk:
392 case sdio0_clk ... spi1_clk:
393 return zynq_clk_get_peripheral_rate(priv, id, two_divs);
394 case dma_clk:
395 return zynq_clk_get_cpu_rate(priv, cpu_2x_clk);
Michal Simek58afff42018-02-21 15:06:20 +0100396 case usb0_aper_clk ... swdt_clk:
Stefan Herbrechtsmeier3a64b252017-01-17 16:27:29 +0100397 return zynq_clk_get_cpu_rate(priv, cpu_1x_clk);
398 default:
399 return -ENXIO;
400 }
401}
402
403static ulong zynq_clk_set_rate(struct clk *clk, ulong rate)
404{
405 struct zynq_clk_priv *priv = dev_get_priv(clk->dev);
406 enum zynq_clk id = clk->id;
407 bool two_divs = false;
408
409 switch (id) {
410 case gem0_clk ... gem1_clk:
411 return zynq_clk_set_gem_rate(priv, id, rate);
412 case fclk0_clk ... can1_clk:
413 two_divs = true;
414 /* fall through */
415 case lqspi_clk ... pcap_clk:
416 case sdio0_clk ... spi1_clk:
417 case dbg_trc_clk ... dbg_apb_clk:
418 return zynq_clk_set_peripheral_rate(priv, id, rate, two_divs);
419 default:
420 return -ENXIO;
421 }
422}
423#else
424static ulong zynq_clk_get_rate(struct clk *clk)
425{
426 struct zynq_clk_priv *priv = dev_get_priv(clk->dev);
427 enum zynq_clk id = clk->id;
428
429 switch (id) {
430 case cpu_6or4x_clk ... cpu_1x_clk:
431 return zynq_clk_get_cpu_rate(priv, id);
432 case ddr3x_clk:
433 return zynq_clk_get_ddr3x_rate(priv);
434 case lqspi_clk ... pcap_clk:
435 case sdio0_clk ... spi1_clk:
436 return zynq_clk_get_peripheral_rate(priv, id, 0);
437 default:
438 return -ENXIO;
439 }
440}
441#endif
442
443static struct clk_ops zynq_clk_ops = {
444 .get_rate = zynq_clk_get_rate,
445#ifndef CONFIG_SPL_BUILD
446 .set_rate = zynq_clk_set_rate,
447#endif
448};
449
450static int zynq_clk_probe(struct udevice *dev)
451{
452 struct zynq_clk_priv *priv = dev_get_priv(dev);
Stefan Herbrechtsmeier9bb803d2017-01-17 16:27:31 +0100453#ifndef CONFIG_SPL_BUILD
454 unsigned int i;
455 char name[16];
456 int ret;
457
458 for (i = 0; i < 2; i++) {
459 sprintf(name, "gem%d_emio_clk", i);
460 ret = clk_get_by_name(dev, name, &priv->gem_emio_clk[i]);
Simon Glassaa9bb092017-05-30 21:47:29 -0600461 if (ret < 0 && ret != -ENODATA) {
Stefan Herbrechtsmeier9bb803d2017-01-17 16:27:31 +0100462 dev_err(dev, "failed to get %s clock\n", name);
463 return ret;
464 }
465 }
466#endif
Stefan Herbrechtsmeier3a64b252017-01-17 16:27:29 +0100467
Simon Glassda409cc2017-05-17 17:18:09 -0600468 priv->ps_clk_freq = fdtdec_get_uint(gd->fdt_blob, dev_of_offset(dev),
Stefan Herbrechtsmeier3a64b252017-01-17 16:27:29 +0100469 "ps-clk-frequency", 33333333UL);
470
471 return 0;
472}
473
474static const struct udevice_id zynq_clk_ids[] = {
475 { .compatible = "xlnx,ps7-clkc"},
476 {}
477};
478
479U_BOOT_DRIVER(zynq_clk) = {
480 .name = "zynq_clk",
481 .id = UCLASS_CLK,
482 .of_match = zynq_clk_ids,
Stefan Herbrechtsmeier3a64b252017-01-17 16:27:29 +0100483 .ops = &zynq_clk_ops,
484 .priv_auto_alloc_size = sizeof(struct zynq_clk_priv),
485 .probe = zynq_clk_probe,
486};