Commit | Line | Data |
---|---|---|
d2912cb1 | 1 | // SPDX-License-Identifier: GPL-2.0-only |
335a1275 LW |
2 | /* |
3 | * Qualcomm External Bus Interface 2 (EBI2) driver | |
4 | * an older version of the Qualcomm Parallel Interface Controller (QPIC) | |
5 | * | |
6 | * Copyright (C) 2016 Linaro Ltd. | |
7 | * | |
8 | * Author: Linus Walleij <linus.walleij@linaro.org> | |
9 | * | |
335a1275 LW |
10 | * See the device tree bindings for this block for more details on the |
11 | * hardware. | |
12 | */ | |
13 | ||
14 | #include <linux/module.h> | |
15 | #include <linux/clk.h> | |
16 | #include <linux/err.h> | |
17 | #include <linux/io.h> | |
18 | #include <linux/of.h> | |
19 | #include <linux/of_platform.h> | |
20 | #include <linux/init.h> | |
335a1275 LW |
21 | #include <linux/slab.h> |
22 | #include <linux/platform_device.h> | |
23 | #include <linux/bitops.h> | |
24 | ||
25 | /* | |
26 | * CS0, CS1, CS4 and CS5 are two bits wide, CS2 and CS3 are one bit. | |
27 | */ | |
28 | #define EBI2_CS0_ENABLE_MASK BIT(0)|BIT(1) | |
29 | #define EBI2_CS1_ENABLE_MASK BIT(2)|BIT(3) | |
30 | #define EBI2_CS2_ENABLE_MASK BIT(4) | |
31 | #define EBI2_CS3_ENABLE_MASK BIT(5) | |
32 | #define EBI2_CS4_ENABLE_MASK BIT(6)|BIT(7) | |
33 | #define EBI2_CS5_ENABLE_MASK BIT(8)|BIT(9) | |
34 | #define EBI2_CSN_MASK GENMASK(9, 0) | |
35 | ||
36 | #define EBI2_XMEM_CFG 0x0000 /* Power management etc */ | |
37 | ||
38 | /* | |
39 | * SLOW CSn CFG | |
40 | * | |
41 | * Bits 31-28: RECOVERY recovery cycles (0 = 1, 1 = 2 etc) this is the time the | |
42 | * memory continues to drive the data bus after OE is de-asserted. | |
43 | * Inserted when reading one CS and switching to another CS or read | |
44 | * followed by write on the same CS. Valid values 0 thru 15. | |
45 | * Bits 27-24: WR_HOLD write hold cycles, these are extra cycles inserted after | |
46 | * every write minimum 1. The data out is driven from the time WE is | |
47 | * asserted until CS is asserted. With a hold of 1, the CS stays | |
48 | * active for 1 extra cycle etc. Valid values 0 thru 15. | |
49 | * Bits 23-16: WR_DELTA initial latency for write cycles inserted for the first | |
50 | * write to a page or burst memory | |
51 | * Bits 15-8: RD_DELTA initial latency for read cycles inserted for the first | |
52 | * read to a page or burst memory | |
53 | * Bits 7-4: WR_WAIT number of wait cycles for every write access, 0=1 cycle | |
54 | * so 1 thru 16 cycles. | |
55 | * Bits 3-0: RD_WAIT number of wait cycles for every read access, 0=1 cycle | |
56 | * so 1 thru 16 cycles. | |
57 | */ | |
58 | #define EBI2_XMEM_CS0_SLOW_CFG 0x0008 | |
59 | #define EBI2_XMEM_CS1_SLOW_CFG 0x000C | |
60 | #define EBI2_XMEM_CS2_SLOW_CFG 0x0010 | |
61 | #define EBI2_XMEM_CS3_SLOW_CFG 0x0014 | |
62 | #define EBI2_XMEM_CS4_SLOW_CFG 0x0018 | |
63 | #define EBI2_XMEM_CS5_SLOW_CFG 0x001C | |
64 | ||
65 | #define EBI2_XMEM_RECOVERY_SHIFT 28 | |
66 | #define EBI2_XMEM_WR_HOLD_SHIFT 24 | |
67 | #define EBI2_XMEM_WR_DELTA_SHIFT 16 | |
68 | #define EBI2_XMEM_RD_DELTA_SHIFT 8 | |
69 | #define EBI2_XMEM_WR_WAIT_SHIFT 4 | |
70 | #define EBI2_XMEM_RD_WAIT_SHIFT 0 | |
71 | ||
72 | /* | |
73 | * FAST CSn CFG | |
74 | * Bits 31-28: ? | |
75 | * Bits 27-24: RD_HOLD: the length in cycles of the first segment of a read | |
76 | * transfer. For a single read trandfer this will be the time | |
77 | * from CS assertion to OE assertion. | |
78 | * Bits 18-24: ? | |
79 | * Bits 17-16: ADV_OE_RECOVERY, the number of cycles elapsed before an OE | |
80 | * assertion, with respect to the cycle where ADV is asserted. | |
81 | * 2 means 2 cycles between ADV and OE. Values 0, 1, 2 or 3. | |
82 | * Bits 5: ADDR_HOLD_ENA, The address is held for an extra cycle to meet | |
83 | * hold time requirements with ADV assertion. | |
84 | * | |
85 | * The manual mentions "write precharge cycles" and "precharge cycles". | |
86 | * We have not been able to figure out which bit fields these correspond to | |
87 | * in the hardware, or what valid values exist. The current hypothesis is that | |
88 | * this is something just used on the FAST chip selects. There is also a "byte | |
89 | * device enable" flag somewhere for 8bit memories. | |
90 | */ | |
91 | #define EBI2_XMEM_CS0_FAST_CFG 0x0028 | |
92 | #define EBI2_XMEM_CS1_FAST_CFG 0x002C | |
93 | #define EBI2_XMEM_CS2_FAST_CFG 0x0030 | |
94 | #define EBI2_XMEM_CS3_FAST_CFG 0x0034 | |
95 | #define EBI2_XMEM_CS4_FAST_CFG 0x0038 | |
96 | #define EBI2_XMEM_CS5_FAST_CFG 0x003C | |
97 | ||
98 | #define EBI2_XMEM_RD_HOLD_SHIFT 24 | |
99 | #define EBI2_XMEM_ADV_OE_RECOVERY_SHIFT 16 | |
100 | #define EBI2_XMEM_ADDR_HOLD_ENA_SHIFT 5 | |
101 | ||
102 | /** | |
103 | * struct cs_data - struct with info on a chipselect setting | |
104 | * @enable_mask: mask to enable the chipselect in the EBI2 config | |
ee704ebf LJ |
105 | * @slow_cfg: offset to XMEMC slow CS config |
106 | * @fast_cfg: offset to XMEMC fast CS config | |
335a1275 LW |
107 | */ |
108 | struct cs_data { | |
109 | u32 enable_mask; | |
110 | u16 slow_cfg; | |
111 | u16 fast_cfg; | |
112 | }; | |
113 | ||
114 | static const struct cs_data cs_info[] = { | |
115 | { | |
116 | /* CS0 */ | |
117 | .enable_mask = EBI2_CS0_ENABLE_MASK, | |
118 | .slow_cfg = EBI2_XMEM_CS0_SLOW_CFG, | |
119 | .fast_cfg = EBI2_XMEM_CS0_FAST_CFG, | |
120 | }, | |
121 | { | |
122 | /* CS1 */ | |
123 | .enable_mask = EBI2_CS1_ENABLE_MASK, | |
124 | .slow_cfg = EBI2_XMEM_CS1_SLOW_CFG, | |
125 | .fast_cfg = EBI2_XMEM_CS1_FAST_CFG, | |
126 | }, | |
127 | { | |
128 | /* CS2 */ | |
129 | .enable_mask = EBI2_CS2_ENABLE_MASK, | |
130 | .slow_cfg = EBI2_XMEM_CS2_SLOW_CFG, | |
131 | .fast_cfg = EBI2_XMEM_CS2_FAST_CFG, | |
132 | }, | |
133 | { | |
134 | /* CS3 */ | |
135 | .enable_mask = EBI2_CS3_ENABLE_MASK, | |
136 | .slow_cfg = EBI2_XMEM_CS3_SLOW_CFG, | |
137 | .fast_cfg = EBI2_XMEM_CS3_FAST_CFG, | |
138 | }, | |
139 | { | |
140 | /* CS4 */ | |
141 | .enable_mask = EBI2_CS4_ENABLE_MASK, | |
142 | .slow_cfg = EBI2_XMEM_CS4_SLOW_CFG, | |
143 | .fast_cfg = EBI2_XMEM_CS4_FAST_CFG, | |
144 | }, | |
145 | { | |
146 | /* CS5 */ | |
147 | .enable_mask = EBI2_CS5_ENABLE_MASK, | |
148 | .slow_cfg = EBI2_XMEM_CS5_SLOW_CFG, | |
149 | .fast_cfg = EBI2_XMEM_CS5_FAST_CFG, | |
150 | }, | |
151 | }; | |
152 | ||
153 | /** | |
154 | * struct ebi2_xmem_prop - describes an XMEM config property | |
155 | * @prop: the device tree binding name | |
156 | * @max: maximum value for the property | |
157 | * @slowreg: true if this property is in the SLOW CS config register | |
158 | * else it is assumed to be in the FAST config register | |
159 | * @shift: the bit field start in the SLOW or FAST register for this | |
160 | * property | |
161 | */ | |
162 | struct ebi2_xmem_prop { | |
163 | const char *prop; | |
164 | u32 max; | |
165 | bool slowreg; | |
166 | u16 shift; | |
167 | }; | |
168 | ||
169 | static const struct ebi2_xmem_prop xmem_props[] = { | |
170 | { | |
171 | .prop = "qcom,xmem-recovery-cycles", | |
172 | .max = 15, | |
173 | .slowreg = true, | |
174 | .shift = EBI2_XMEM_RECOVERY_SHIFT, | |
175 | }, | |
176 | { | |
177 | .prop = "qcom,xmem-write-hold-cycles", | |
178 | .max = 15, | |
179 | .slowreg = true, | |
180 | .shift = EBI2_XMEM_WR_HOLD_SHIFT, | |
181 | }, | |
182 | { | |
183 | .prop = "qcom,xmem-write-delta-cycles", | |
184 | .max = 255, | |
185 | .slowreg = true, | |
186 | .shift = EBI2_XMEM_WR_DELTA_SHIFT, | |
187 | }, | |
188 | { | |
189 | .prop = "qcom,xmem-read-delta-cycles", | |
190 | .max = 255, | |
191 | .slowreg = true, | |
192 | .shift = EBI2_XMEM_RD_DELTA_SHIFT, | |
193 | }, | |
194 | { | |
195 | .prop = "qcom,xmem-write-wait-cycles", | |
196 | .max = 15, | |
197 | .slowreg = true, | |
198 | .shift = EBI2_XMEM_WR_WAIT_SHIFT, | |
199 | }, | |
200 | { | |
201 | .prop = "qcom,xmem-read-wait-cycles", | |
202 | .max = 15, | |
203 | .slowreg = true, | |
204 | .shift = EBI2_XMEM_RD_WAIT_SHIFT, | |
205 | }, | |
206 | { | |
207 | .prop = "qcom,xmem-address-hold-enable", | |
208 | .max = 1, /* boolean prop */ | |
209 | .slowreg = false, | |
210 | .shift = EBI2_XMEM_ADDR_HOLD_ENA_SHIFT, | |
211 | }, | |
212 | { | |
213 | .prop = "qcom,xmem-adv-to-oe-recovery-cycles", | |
214 | .max = 3, | |
215 | .slowreg = false, | |
216 | .shift = EBI2_XMEM_ADV_OE_RECOVERY_SHIFT, | |
217 | }, | |
218 | { | |
219 | .prop = "qcom,xmem-read-hold-cycles", | |
220 | .max = 15, | |
221 | .slowreg = false, | |
222 | .shift = EBI2_XMEM_RD_HOLD_SHIFT, | |
223 | }, | |
224 | }; | |
225 | ||
226 | static void qcom_ebi2_setup_chipselect(struct device_node *np, | |
227 | struct device *dev, | |
228 | void __iomem *ebi2_base, | |
229 | void __iomem *ebi2_xmem, | |
230 | u32 csindex) | |
231 | { | |
232 | const struct cs_data *csd; | |
233 | u32 slowcfg, fastcfg; | |
234 | u32 val; | |
235 | int ret; | |
236 | int i; | |
237 | ||
238 | csd = &cs_info[csindex]; | |
239 | val = readl(ebi2_base); | |
240 | val |= csd->enable_mask; | |
241 | writel(val, ebi2_base); | |
242 | dev_dbg(dev, "enabled CS%u\n", csindex); | |
243 | ||
244 | /* Next set up the XMEMC */ | |
245 | slowcfg = 0; | |
246 | fastcfg = 0; | |
247 | ||
248 | for (i = 0; i < ARRAY_SIZE(xmem_props); i++) { | |
249 | const struct ebi2_xmem_prop *xp = &xmem_props[i]; | |
250 | ||
251 | /* All are regular u32 values */ | |
252 | ret = of_property_read_u32(np, xp->prop, &val); | |
253 | if (ret) { | |
254 | dev_dbg(dev, "could not read %s for CS%d\n", | |
255 | xp->prop, csindex); | |
256 | continue; | |
257 | } | |
258 | ||
259 | /* First check boolean props */ | |
260 | if (xp->max == 1 && val) { | |
261 | if (xp->slowreg) | |
262 | slowcfg |= BIT(xp->shift); | |
263 | else | |
264 | fastcfg |= BIT(xp->shift); | |
265 | dev_dbg(dev, "set %s flag\n", xp->prop); | |
266 | continue; | |
267 | } | |
268 | ||
269 | /* We're dealing with an u32 */ | |
270 | if (val > xp->max) { | |
271 | dev_err(dev, | |
272 | "too high value for %s: %u, capped at %u\n", | |
273 | xp->prop, val, xp->max); | |
274 | val = xp->max; | |
275 | } | |
276 | if (xp->slowreg) | |
277 | slowcfg |= (val << xp->shift); | |
278 | else | |
279 | fastcfg |= (val << xp->shift); | |
280 | dev_dbg(dev, "set %s to %u\n", xp->prop, val); | |
281 | } | |
282 | ||
283 | dev_info(dev, "CS%u: SLOW CFG 0x%08x, FAST CFG 0x%08x\n", | |
284 | csindex, slowcfg, fastcfg); | |
285 | ||
286 | if (slowcfg) | |
287 | writel(slowcfg, ebi2_xmem + csd->slow_cfg); | |
288 | if (fastcfg) | |
289 | writel(fastcfg, ebi2_xmem + csd->fast_cfg); | |
290 | } | |
291 | ||
292 | static int qcom_ebi2_probe(struct platform_device *pdev) | |
293 | { | |
294 | struct device_node *np = pdev->dev.of_node; | |
295 | struct device_node *child; | |
296 | struct device *dev = &pdev->dev; | |
297 | struct resource *res; | |
298 | void __iomem *ebi2_base; | |
299 | void __iomem *ebi2_xmem; | |
300 | struct clk *ebi2xclk; | |
301 | struct clk *ebi2clk; | |
302 | bool have_children = false; | |
303 | u32 val; | |
304 | int ret; | |
305 | ||
306 | ebi2xclk = devm_clk_get(dev, "ebi2x"); | |
307 | if (IS_ERR(ebi2xclk)) | |
308 | return PTR_ERR(ebi2xclk); | |
309 | ||
310 | ret = clk_prepare_enable(ebi2xclk); | |
311 | if (ret) { | |
312 | dev_err(dev, "could not enable EBI2X clk (%d)\n", ret); | |
313 | return ret; | |
314 | } | |
315 | ||
316 | ebi2clk = devm_clk_get(dev, "ebi2"); | |
317 | if (IS_ERR(ebi2clk)) { | |
318 | ret = PTR_ERR(ebi2clk); | |
319 | goto err_disable_2x_clk; | |
320 | } | |
321 | ||
322 | ret = clk_prepare_enable(ebi2clk); | |
323 | if (ret) { | |
324 | dev_err(dev, "could not enable EBI2 clk\n"); | |
325 | goto err_disable_2x_clk; | |
326 | } | |
327 | ||
328 | res = platform_get_resource(pdev, IORESOURCE_MEM, 0); | |
329 | ebi2_base = devm_ioremap_resource(dev, res); | |
330 | if (IS_ERR(ebi2_base)) { | |
331 | ret = PTR_ERR(ebi2_base); | |
332 | goto err_disable_clk; | |
333 | } | |
334 | ||
335 | res = platform_get_resource(pdev, IORESOURCE_MEM, 1); | |
336 | ebi2_xmem = devm_ioremap_resource(dev, res); | |
337 | if (IS_ERR(ebi2_xmem)) { | |
338 | ret = PTR_ERR(ebi2_xmem); | |
339 | goto err_disable_clk; | |
340 | } | |
341 | ||
342 | /* Allegedly this turns the power save mode off */ | |
343 | writel(0UL, ebi2_xmem + EBI2_XMEM_CFG); | |
344 | ||
345 | /* Disable all chipselects */ | |
346 | val = readl(ebi2_base); | |
347 | val &= ~EBI2_CSN_MASK; | |
348 | writel(val, ebi2_base); | |
349 | ||
350 | /* Walk over the child nodes and see what chipselects we use */ | |
351 | for_each_available_child_of_node(np, child) { | |
352 | u32 csindex; | |
353 | ||
354 | /* Figure out the chipselect */ | |
355 | ret = of_property_read_u32(child, "reg", &csindex); | |
ac6ad7c2 PB |
356 | if (ret) { |
357 | of_node_put(child); | |
335a1275 | 358 | return ret; |
ac6ad7c2 | 359 | } |
335a1275 LW |
360 | |
361 | if (csindex > 5) { | |
362 | dev_err(dev, | |
363 | "invalid chipselect %u, we only support 0-5\n", | |
364 | csindex); | |
365 | continue; | |
366 | } | |
367 | ||
368 | qcom_ebi2_setup_chipselect(child, | |
369 | dev, | |
370 | ebi2_base, | |
371 | ebi2_xmem, | |
372 | csindex); | |
373 | ||
374 | /* We have at least one child */ | |
375 | have_children = true; | |
376 | } | |
377 | ||
378 | if (have_children) | |
379 | return of_platform_default_populate(np, NULL, dev); | |
380 | return 0; | |
381 | ||
382 | err_disable_clk: | |
383 | clk_disable_unprepare(ebi2clk); | |
384 | err_disable_2x_clk: | |
385 | clk_disable_unprepare(ebi2xclk); | |
386 | ||
387 | return ret; | |
388 | } | |
389 | ||
390 | static const struct of_device_id qcom_ebi2_of_match[] = { | |
391 | { .compatible = "qcom,msm8660-ebi2", }, | |
392 | { .compatible = "qcom,apq8060-ebi2", }, | |
393 | { } | |
394 | }; | |
395 | ||
396 | static struct platform_driver qcom_ebi2_driver = { | |
397 | .probe = qcom_ebi2_probe, | |
398 | .driver = { | |
399 | .name = "qcom-ebi2", | |
400 | .of_match_table = qcom_ebi2_of_match, | |
401 | }, | |
402 | }; | |
403 | module_platform_driver(qcom_ebi2_driver); | |
404 | MODULE_AUTHOR("Linus Walleij <linus.walleij@linaro.org>"); | |
405 | MODULE_DESCRIPTION("Qualcomm EBI2 driver"); |