Merge branch 'master' of git://git.kernel.org/pub/scm/linux/kernel/git/klassert/ipsec
[linux-2.6-block.git] / drivers / bus / qcom-ebi2.c
CommitLineData
335a1275
LW
1/*
2 * Qualcomm External Bus Interface 2 (EBI2) driver
3 * an older version of the Qualcomm Parallel Interface Controller (QPIC)
4 *
5 * Copyright (C) 2016 Linaro Ltd.
6 *
7 * Author: Linus Walleij <linus.walleij@linaro.org>
8 *
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of the GNU General Public License version 2, as
11 * published by the Free Software Foundation.
12 *
13 * See the device tree bindings for this block for more details on the
14 * hardware.
15 */
16
17#include <linux/module.h>
18#include <linux/clk.h>
19#include <linux/err.h>
20#include <linux/io.h>
21#include <linux/of.h>
22#include <linux/of_platform.h>
23#include <linux/init.h>
335a1275
LW
24#include <linux/slab.h>
25#include <linux/platform_device.h>
26#include <linux/bitops.h>
27
28/*
29 * CS0, CS1, CS4 and CS5 are two bits wide, CS2 and CS3 are one bit.
30 */
31#define EBI2_CS0_ENABLE_MASK BIT(0)|BIT(1)
32#define EBI2_CS1_ENABLE_MASK BIT(2)|BIT(3)
33#define EBI2_CS2_ENABLE_MASK BIT(4)
34#define EBI2_CS3_ENABLE_MASK BIT(5)
35#define EBI2_CS4_ENABLE_MASK BIT(6)|BIT(7)
36#define EBI2_CS5_ENABLE_MASK BIT(8)|BIT(9)
37#define EBI2_CSN_MASK GENMASK(9, 0)
38
39#define EBI2_XMEM_CFG 0x0000 /* Power management etc */
40
41/*
42 * SLOW CSn CFG
43 *
44 * Bits 31-28: RECOVERY recovery cycles (0 = 1, 1 = 2 etc) this is the time the
45 * memory continues to drive the data bus after OE is de-asserted.
46 * Inserted when reading one CS and switching to another CS or read
47 * followed by write on the same CS. Valid values 0 thru 15.
48 * Bits 27-24: WR_HOLD write hold cycles, these are extra cycles inserted after
49 * every write minimum 1. The data out is driven from the time WE is
50 * asserted until CS is asserted. With a hold of 1, the CS stays
51 * active for 1 extra cycle etc. Valid values 0 thru 15.
52 * Bits 23-16: WR_DELTA initial latency for write cycles inserted for the first
53 * write to a page or burst memory
54 * Bits 15-8: RD_DELTA initial latency for read cycles inserted for the first
55 * read to a page or burst memory
56 * Bits 7-4: WR_WAIT number of wait cycles for every write access, 0=1 cycle
57 * so 1 thru 16 cycles.
58 * Bits 3-0: RD_WAIT number of wait cycles for every read access, 0=1 cycle
59 * so 1 thru 16 cycles.
60 */
61#define EBI2_XMEM_CS0_SLOW_CFG 0x0008
62#define EBI2_XMEM_CS1_SLOW_CFG 0x000C
63#define EBI2_XMEM_CS2_SLOW_CFG 0x0010
64#define EBI2_XMEM_CS3_SLOW_CFG 0x0014
65#define EBI2_XMEM_CS4_SLOW_CFG 0x0018
66#define EBI2_XMEM_CS5_SLOW_CFG 0x001C
67
68#define EBI2_XMEM_RECOVERY_SHIFT 28
69#define EBI2_XMEM_WR_HOLD_SHIFT 24
70#define EBI2_XMEM_WR_DELTA_SHIFT 16
71#define EBI2_XMEM_RD_DELTA_SHIFT 8
72#define EBI2_XMEM_WR_WAIT_SHIFT 4
73#define EBI2_XMEM_RD_WAIT_SHIFT 0
74
75/*
76 * FAST CSn CFG
77 * Bits 31-28: ?
78 * Bits 27-24: RD_HOLD: the length in cycles of the first segment of a read
79 * transfer. For a single read trandfer this will be the time
80 * from CS assertion to OE assertion.
81 * Bits 18-24: ?
82 * Bits 17-16: ADV_OE_RECOVERY, the number of cycles elapsed before an OE
83 * assertion, with respect to the cycle where ADV is asserted.
84 * 2 means 2 cycles between ADV and OE. Values 0, 1, 2 or 3.
85 * Bits 5: ADDR_HOLD_ENA, The address is held for an extra cycle to meet
86 * hold time requirements with ADV assertion.
87 *
88 * The manual mentions "write precharge cycles" and "precharge cycles".
89 * We have not been able to figure out which bit fields these correspond to
90 * in the hardware, or what valid values exist. The current hypothesis is that
91 * this is something just used on the FAST chip selects. There is also a "byte
92 * device enable" flag somewhere for 8bit memories.
93 */
94#define EBI2_XMEM_CS0_FAST_CFG 0x0028
95#define EBI2_XMEM_CS1_FAST_CFG 0x002C
96#define EBI2_XMEM_CS2_FAST_CFG 0x0030
97#define EBI2_XMEM_CS3_FAST_CFG 0x0034
98#define EBI2_XMEM_CS4_FAST_CFG 0x0038
99#define EBI2_XMEM_CS5_FAST_CFG 0x003C
100
101#define EBI2_XMEM_RD_HOLD_SHIFT 24
102#define EBI2_XMEM_ADV_OE_RECOVERY_SHIFT 16
103#define EBI2_XMEM_ADDR_HOLD_ENA_SHIFT 5
104
105/**
106 * struct cs_data - struct with info on a chipselect setting
107 * @enable_mask: mask to enable the chipselect in the EBI2 config
108 * @slow_cfg0: offset to XMEMC slow CS config
109 * @fast_cfg1: offset to XMEMC fast CS config
110 */
111struct cs_data {
112 u32 enable_mask;
113 u16 slow_cfg;
114 u16 fast_cfg;
115};
116
117static const struct cs_data cs_info[] = {
118 {
119 /* CS0 */
120 .enable_mask = EBI2_CS0_ENABLE_MASK,
121 .slow_cfg = EBI2_XMEM_CS0_SLOW_CFG,
122 .fast_cfg = EBI2_XMEM_CS0_FAST_CFG,
123 },
124 {
125 /* CS1 */
126 .enable_mask = EBI2_CS1_ENABLE_MASK,
127 .slow_cfg = EBI2_XMEM_CS1_SLOW_CFG,
128 .fast_cfg = EBI2_XMEM_CS1_FAST_CFG,
129 },
130 {
131 /* CS2 */
132 .enable_mask = EBI2_CS2_ENABLE_MASK,
133 .slow_cfg = EBI2_XMEM_CS2_SLOW_CFG,
134 .fast_cfg = EBI2_XMEM_CS2_FAST_CFG,
135 },
136 {
137 /* CS3 */
138 .enable_mask = EBI2_CS3_ENABLE_MASK,
139 .slow_cfg = EBI2_XMEM_CS3_SLOW_CFG,
140 .fast_cfg = EBI2_XMEM_CS3_FAST_CFG,
141 },
142 {
143 /* CS4 */
144 .enable_mask = EBI2_CS4_ENABLE_MASK,
145 .slow_cfg = EBI2_XMEM_CS4_SLOW_CFG,
146 .fast_cfg = EBI2_XMEM_CS4_FAST_CFG,
147 },
148 {
149 /* CS5 */
150 .enable_mask = EBI2_CS5_ENABLE_MASK,
151 .slow_cfg = EBI2_XMEM_CS5_SLOW_CFG,
152 .fast_cfg = EBI2_XMEM_CS5_FAST_CFG,
153 },
154};
155
156/**
157 * struct ebi2_xmem_prop - describes an XMEM config property
158 * @prop: the device tree binding name
159 * @max: maximum value for the property
160 * @slowreg: true if this property is in the SLOW CS config register
161 * else it is assumed to be in the FAST config register
162 * @shift: the bit field start in the SLOW or FAST register for this
163 * property
164 */
165struct ebi2_xmem_prop {
166 const char *prop;
167 u32 max;
168 bool slowreg;
169 u16 shift;
170};
171
172static const struct ebi2_xmem_prop xmem_props[] = {
173 {
174 .prop = "qcom,xmem-recovery-cycles",
175 .max = 15,
176 .slowreg = true,
177 .shift = EBI2_XMEM_RECOVERY_SHIFT,
178 },
179 {
180 .prop = "qcom,xmem-write-hold-cycles",
181 .max = 15,
182 .slowreg = true,
183 .shift = EBI2_XMEM_WR_HOLD_SHIFT,
184 },
185 {
186 .prop = "qcom,xmem-write-delta-cycles",
187 .max = 255,
188 .slowreg = true,
189 .shift = EBI2_XMEM_WR_DELTA_SHIFT,
190 },
191 {
192 .prop = "qcom,xmem-read-delta-cycles",
193 .max = 255,
194 .slowreg = true,
195 .shift = EBI2_XMEM_RD_DELTA_SHIFT,
196 },
197 {
198 .prop = "qcom,xmem-write-wait-cycles",
199 .max = 15,
200 .slowreg = true,
201 .shift = EBI2_XMEM_WR_WAIT_SHIFT,
202 },
203 {
204 .prop = "qcom,xmem-read-wait-cycles",
205 .max = 15,
206 .slowreg = true,
207 .shift = EBI2_XMEM_RD_WAIT_SHIFT,
208 },
209 {
210 .prop = "qcom,xmem-address-hold-enable",
211 .max = 1, /* boolean prop */
212 .slowreg = false,
213 .shift = EBI2_XMEM_ADDR_HOLD_ENA_SHIFT,
214 },
215 {
216 .prop = "qcom,xmem-adv-to-oe-recovery-cycles",
217 .max = 3,
218 .slowreg = false,
219 .shift = EBI2_XMEM_ADV_OE_RECOVERY_SHIFT,
220 },
221 {
222 .prop = "qcom,xmem-read-hold-cycles",
223 .max = 15,
224 .slowreg = false,
225 .shift = EBI2_XMEM_RD_HOLD_SHIFT,
226 },
227};
228
229static void qcom_ebi2_setup_chipselect(struct device_node *np,
230 struct device *dev,
231 void __iomem *ebi2_base,
232 void __iomem *ebi2_xmem,
233 u32 csindex)
234{
235 const struct cs_data *csd;
236 u32 slowcfg, fastcfg;
237 u32 val;
238 int ret;
239 int i;
240
241 csd = &cs_info[csindex];
242 val = readl(ebi2_base);
243 val |= csd->enable_mask;
244 writel(val, ebi2_base);
245 dev_dbg(dev, "enabled CS%u\n", csindex);
246
247 /* Next set up the XMEMC */
248 slowcfg = 0;
249 fastcfg = 0;
250
251 for (i = 0; i < ARRAY_SIZE(xmem_props); i++) {
252 const struct ebi2_xmem_prop *xp = &xmem_props[i];
253
254 /* All are regular u32 values */
255 ret = of_property_read_u32(np, xp->prop, &val);
256 if (ret) {
257 dev_dbg(dev, "could not read %s for CS%d\n",
258 xp->prop, csindex);
259 continue;
260 }
261
262 /* First check boolean props */
263 if (xp->max == 1 && val) {
264 if (xp->slowreg)
265 slowcfg |= BIT(xp->shift);
266 else
267 fastcfg |= BIT(xp->shift);
268 dev_dbg(dev, "set %s flag\n", xp->prop);
269 continue;
270 }
271
272 /* We're dealing with an u32 */
273 if (val > xp->max) {
274 dev_err(dev,
275 "too high value for %s: %u, capped at %u\n",
276 xp->prop, val, xp->max);
277 val = xp->max;
278 }
279 if (xp->slowreg)
280 slowcfg |= (val << xp->shift);
281 else
282 fastcfg |= (val << xp->shift);
283 dev_dbg(dev, "set %s to %u\n", xp->prop, val);
284 }
285
286 dev_info(dev, "CS%u: SLOW CFG 0x%08x, FAST CFG 0x%08x\n",
287 csindex, slowcfg, fastcfg);
288
289 if (slowcfg)
290 writel(slowcfg, ebi2_xmem + csd->slow_cfg);
291 if (fastcfg)
292 writel(fastcfg, ebi2_xmem + csd->fast_cfg);
293}
294
295static int qcom_ebi2_probe(struct platform_device *pdev)
296{
297 struct device_node *np = pdev->dev.of_node;
298 struct device_node *child;
299 struct device *dev = &pdev->dev;
300 struct resource *res;
301 void __iomem *ebi2_base;
302 void __iomem *ebi2_xmem;
303 struct clk *ebi2xclk;
304 struct clk *ebi2clk;
305 bool have_children = false;
306 u32 val;
307 int ret;
308
309 ebi2xclk = devm_clk_get(dev, "ebi2x");
310 if (IS_ERR(ebi2xclk))
311 return PTR_ERR(ebi2xclk);
312
313 ret = clk_prepare_enable(ebi2xclk);
314 if (ret) {
315 dev_err(dev, "could not enable EBI2X clk (%d)\n", ret);
316 return ret;
317 }
318
319 ebi2clk = devm_clk_get(dev, "ebi2");
320 if (IS_ERR(ebi2clk)) {
321 ret = PTR_ERR(ebi2clk);
322 goto err_disable_2x_clk;
323 }
324
325 ret = clk_prepare_enable(ebi2clk);
326 if (ret) {
327 dev_err(dev, "could not enable EBI2 clk\n");
328 goto err_disable_2x_clk;
329 }
330
331 res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
332 ebi2_base = devm_ioremap_resource(dev, res);
333 if (IS_ERR(ebi2_base)) {
334 ret = PTR_ERR(ebi2_base);
335 goto err_disable_clk;
336 }
337
338 res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
339 ebi2_xmem = devm_ioremap_resource(dev, res);
340 if (IS_ERR(ebi2_xmem)) {
341 ret = PTR_ERR(ebi2_xmem);
342 goto err_disable_clk;
343 }
344
345 /* Allegedly this turns the power save mode off */
346 writel(0UL, ebi2_xmem + EBI2_XMEM_CFG);
347
348 /* Disable all chipselects */
349 val = readl(ebi2_base);
350 val &= ~EBI2_CSN_MASK;
351 writel(val, ebi2_base);
352
353 /* Walk over the child nodes and see what chipselects we use */
354 for_each_available_child_of_node(np, child) {
355 u32 csindex;
356
357 /* Figure out the chipselect */
358 ret = of_property_read_u32(child, "reg", &csindex);
359 if (ret)
360 return ret;
361
362 if (csindex > 5) {
363 dev_err(dev,
364 "invalid chipselect %u, we only support 0-5\n",
365 csindex);
366 continue;
367 }
368
369 qcom_ebi2_setup_chipselect(child,
370 dev,
371 ebi2_base,
372 ebi2_xmem,
373 csindex);
374
375 /* We have at least one child */
376 have_children = true;
377 }
378
379 if (have_children)
380 return of_platform_default_populate(np, NULL, dev);
381 return 0;
382
383err_disable_clk:
384 clk_disable_unprepare(ebi2clk);
385err_disable_2x_clk:
386 clk_disable_unprepare(ebi2xclk);
387
388 return ret;
389}
390
391static const struct of_device_id qcom_ebi2_of_match[] = {
392 { .compatible = "qcom,msm8660-ebi2", },
393 { .compatible = "qcom,apq8060-ebi2", },
394 { }
395};
396
397static struct platform_driver qcom_ebi2_driver = {
398 .probe = qcom_ebi2_probe,
399 .driver = {
400 .name = "qcom-ebi2",
401 .of_match_table = qcom_ebi2_of_match,
402 },
403};
404module_platform_driver(qcom_ebi2_driver);
405MODULE_AUTHOR("Linus Walleij <linus.walleij@linaro.org>");
406MODULE_DESCRIPTION("Qualcomm EBI2 driver");
407MODULE_LICENSE("GPL");