Merge tag 'mvebu-fixes-5.2-2' of git://git.infradead.org/linux-mvebu into arm/fixes
[linux-2.6-block.git] / arch / powerpc / kernel / swsusp_32.S
CommitLineData
b2441318 1/* SPDX-License-Identifier: GPL-2.0 */
1da177e4
LT
2#include <linux/threads.h>
3#include <asm/processor.h>
4#include <asm/page.h>
5#include <asm/cputable.h>
6#include <asm/thread_info.h>
7#include <asm/ppc_asm.h>
0013a854 8#include <asm/asm-offsets.h>
7c03d653 9#include <asm/mmu.h>
2c86cd18 10#include <asm/feature-fixups.h>
1da177e4
LT
11
12/*
13 * Structure for storing CPU registers on the save area.
14 */
15#define SL_SP 0
16#define SL_PC 4
17#define SL_MSR 8
18#define SL_SDR1 0xc
19#define SL_SPRG0 0x10 /* 4 sprg's */
20#define SL_DBAT0 0x20
21#define SL_IBAT0 0x28
22#define SL_DBAT1 0x30
23#define SL_IBAT1 0x38
24#define SL_DBAT2 0x40
25#define SL_IBAT2 0x48
26#define SL_DBAT3 0x50
27#define SL_IBAT3 0x58
28#define SL_TB 0x60
29#define SL_R2 0x68
30#define SL_CR 0x6c
31#define SL_LR 0x70
32#define SL_R12 0x74 /* r12 to r31 */
33#define SL_SIZE (SL_R12 + 80)
34
35 .section .data
36 .align 5
37
38_GLOBAL(swsusp_save_area)
39 .space SL_SIZE
40
41
42 .section .text
43 .align 5
44
45_GLOBAL(swsusp_arch_suspend)
46
47 lis r11,swsusp_save_area@h
48 ori r11,r11,swsusp_save_area@l
49
50 mflr r0
51 stw r0,SL_LR(r11)
52 mfcr r0
53 stw r0,SL_CR(r11)
54 stw r1,SL_SP(r11)
55 stw r2,SL_R2(r11)
56 stmw r12,SL_R12(r11)
57
58 /* Save MSR & SDR1 */
59 mfmsr r4
60 stw r4,SL_MSR(r11)
61 mfsdr1 r4
62 stw r4,SL_SDR1(r11)
63
64 /* Get a stable timebase and save it */
651: mftbu r4
66 stw r4,SL_TB(r11)
67 mftb r5
68 stw r5,SL_TB+4(r11)
69 mftbu r3
70 cmpw r3,r4
71 bne 1b
72
73 /* Save SPRGs */
74 mfsprg r4,0
75 stw r4,SL_SPRG0(r11)
76 mfsprg r4,1
77 stw r4,SL_SPRG0+4(r11)
78 mfsprg r4,2
79 stw r4,SL_SPRG0+8(r11)
80 mfsprg r4,3
81 stw r4,SL_SPRG0+12(r11)
82
83 /* Save BATs */
84 mfdbatu r4,0
85 stw r4,SL_DBAT0(r11)
86 mfdbatl r4,0
87 stw r4,SL_DBAT0+4(r11)
88 mfdbatu r4,1
89 stw r4,SL_DBAT1(r11)
90 mfdbatl r4,1
91 stw r4,SL_DBAT1+4(r11)
92 mfdbatu r4,2
93 stw r4,SL_DBAT2(r11)
94 mfdbatl r4,2
95 stw r4,SL_DBAT2+4(r11)
96 mfdbatu r4,3
97 stw r4,SL_DBAT3(r11)
98 mfdbatl r4,3
99 stw r4,SL_DBAT3+4(r11)
100 mfibatu r4,0
101 stw r4,SL_IBAT0(r11)
102 mfibatl r4,0
103 stw r4,SL_IBAT0+4(r11)
104 mfibatu r4,1
105 stw r4,SL_IBAT1(r11)
106 mfibatl r4,1
107 stw r4,SL_IBAT1+4(r11)
108 mfibatu r4,2
109 stw r4,SL_IBAT2(r11)
110 mfibatl r4,2
111 stw r4,SL_IBAT2+4(r11)
112 mfibatu r4,3
113 stw r4,SL_IBAT3(r11)
114 mfibatl r4,3
115 stw r4,SL_IBAT3+4(r11)
116
117#if 0
118 /* Backup various CPU config stuffs */
119 bl __save_cpu_setup
120#endif
121 /* Call the low level suspend stuff (we should probably have made
122 * a stackframe...
123 */
124 bl swsusp_save
125
126 /* Restore LR from the save area */
127 lis r11,swsusp_save_area@h
128 ori r11,r11,swsusp_save_area@l
129 lwz r0,SL_LR(r11)
130 mtlr r0
131
132 blr
133
134
135/* Resume code */
136_GLOBAL(swsusp_arch_resume)
137
2e6f40de 138#ifdef CONFIG_ALTIVEC
1da177e4
LT
139 /* Stop pending alitvec streams and memory accesses */
140BEGIN_FTR_SECTION
141 DSSALL
142END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC)
2e6f40de 143#endif
1da177e4
LT
144 sync
145
146 /* Disable MSR:DR to make sure we don't take a TLB or
147 * hash miss during the copy, as our hash table will
25985edc 148 * for a while be unusable. For .text, we assume we are
1da177e4
LT
149 * covered by a BAT. This works only for non-G5 at this
150 * point. G5 will need a better approach, possibly using
151 * a small temporary hash table filled with large mappings,
152 * disabling the MMU completely isn't a good option for
153 * performance reasons.
154 * (Note that 750's may have the same performance issue as
155 * the G5 in this case, we should investigate using moving
156 * BATs for these CPUs)
157 */
158 mfmsr r0
159 sync
160 rlwinm r0,r0,0,28,26 /* clear MSR_DR */
161 mtmsr r0
162 sync
163 isync
164
165 /* Load ptr the list of pages to copy in r3 */
75534b50
RW
166 lis r11,(restore_pblist - KERNELBASE)@h
167 ori r11,r11,restore_pblist@l
1da177e4
LT
168 lwz r10,0(r11)
169
170 /* Copy the pages. This is a very basic implementation, to
171 * be replaced by something more cache efficient */
1721:
173 tophys(r3,r10)
174 li r0,256
175 mtctr r0
176 lwz r11,pbe_address(r3) /* source */
177 tophys(r5,r11)
178 lwz r10,pbe_orig_address(r3) /* destination */
179 tophys(r6,r10)
1802:
181 lwz r8,0(r5)
182 lwz r9,4(r5)
183 lwz r10,8(r5)
184 lwz r11,12(r5)
185 addi r5,r5,16
186 stw r8,0(r6)
187 stw r9,4(r6)
188 stw r10,8(r6)
189 stw r11,12(r6)
190 addi r6,r6,16
191 bdnz 2b
192 lwz r10,pbe_next(r3)
193 cmpwi 0,r10,0
194 bne 1b
195
196 /* Do a very simple cache flush/inval of the L1 to ensure
197 * coherency of the icache
198 */
199 lis r3,0x0002
200 mtctr r3
201 li r3, 0
2021:
203 lwz r0,0(r3)
204 addi r3,r3,0x0020
205 bdnz 1b
206 isync
207 sync
208
209 /* Now flush those cache lines */
210 lis r3,0x0002
211 mtctr r3
212 li r3, 0
2131:
214 dcbf 0,r3
215 addi r3,r3,0x0020
216 bdnz 1b
217 sync
218
219 /* Ok, we are now running with the kernel data of the old
220 * kernel fully restored. We can get to the save area
221 * easily now. As for the rest of the code, it assumes the
222 * loader kernel and the booted one are exactly identical
223 */
224 lis r11,swsusp_save_area@h
225 ori r11,r11,swsusp_save_area@l
226 tophys(r11,r11)
227
228#if 0
229 /* Restore various CPU config stuffs */
230 bl __restore_cpu_setup
231#endif
232 /* Restore the BATs, and SDR1. Then we can turn on the MMU.
233 * This is a bit hairy as we are running out of those BATs,
234 * but first, our code is probably in the icache, and we are
235 * writing the same value to the BAT, so that should be fine,
236 * though a better solution will have to be found long-term
237 */
238 lwz r4,SL_SDR1(r11)
239 mtsdr1 r4
240 lwz r4,SL_SPRG0(r11)
241 mtsprg 0,r4
242 lwz r4,SL_SPRG0+4(r11)
243 mtsprg 1,r4
244 lwz r4,SL_SPRG0+8(r11)
245 mtsprg 2,r4
246 lwz r4,SL_SPRG0+12(r11)
247 mtsprg 3,r4
248
249#if 0
250 lwz r4,SL_DBAT0(r11)
251 mtdbatu 0,r4
252 lwz r4,SL_DBAT0+4(r11)
253 mtdbatl 0,r4
254 lwz r4,SL_DBAT1(r11)
255 mtdbatu 1,r4
256 lwz r4,SL_DBAT1+4(r11)
257 mtdbatl 1,r4
258 lwz r4,SL_DBAT2(r11)
259 mtdbatu 2,r4
260 lwz r4,SL_DBAT2+4(r11)
261 mtdbatl 2,r4
262 lwz r4,SL_DBAT3(r11)
263 mtdbatu 3,r4
264 lwz r4,SL_DBAT3+4(r11)
265 mtdbatl 3,r4
266 lwz r4,SL_IBAT0(r11)
267 mtibatu 0,r4
268 lwz r4,SL_IBAT0+4(r11)
269 mtibatl 0,r4
270 lwz r4,SL_IBAT1(r11)
271 mtibatu 1,r4
272 lwz r4,SL_IBAT1+4(r11)
273 mtibatl 1,r4
274 lwz r4,SL_IBAT2(r11)
275 mtibatu 2,r4
276 lwz r4,SL_IBAT2+4(r11)
277 mtibatl 2,r4
278 lwz r4,SL_IBAT3(r11)
279 mtibatu 3,r4
280 lwz r4,SL_IBAT3+4(r11)
281 mtibatl 3,r4
282#endif
283
7c03d653 284BEGIN_MMU_FTR_SECTION
1da177e4
LT
285 li r4,0
286 mtspr SPRN_DBAT4U,r4
287 mtspr SPRN_DBAT4L,r4
288 mtspr SPRN_DBAT5U,r4
289 mtspr SPRN_DBAT5L,r4
290 mtspr SPRN_DBAT6U,r4
291 mtspr SPRN_DBAT6L,r4
292 mtspr SPRN_DBAT7U,r4
293 mtspr SPRN_DBAT7L,r4
294 mtspr SPRN_IBAT4U,r4
295 mtspr SPRN_IBAT4L,r4
296 mtspr SPRN_IBAT5U,r4
297 mtspr SPRN_IBAT5L,r4
298 mtspr SPRN_IBAT6U,r4
299 mtspr SPRN_IBAT6L,r4
300 mtspr SPRN_IBAT7U,r4
301 mtspr SPRN_IBAT7L,r4
7c03d653 302END_MMU_FTR_SECTION_IFSET(MMU_FTR_USE_HIGH_BATS)
1da177e4
LT
303
304 /* Flush all TLBs */
305 lis r4,0x1000
3061: addic. r4,r4,-0x1000
307 tlbie r4
e443ed35 308 bgt 1b
1da177e4
LT
309 sync
310
311 /* restore the MSR and turn on the MMU */
312 lwz r3,SL_MSR(r11)
313 bl turn_on_mmu
314 tovirt(r11,r11)
315
316 /* Restore TB */
317 li r3,0
318 mttbl r3
319 lwz r3,SL_TB(r11)
320 lwz r4,SL_TB+4(r11)
321 mttbu r3
322 mttbl r4
323
324 /* Kick decrementer */
325 li r0,1
326 mtdec r0
327
328 /* Restore the callee-saved registers and return */
329 lwz r0,SL_CR(r11)
330 mtcr r0
331 lwz r2,SL_R2(r11)
332 lmw r12,SL_R12(r11)
333 lwz r1,SL_SP(r11)
334 lwz r0,SL_LR(r11)
335 mtlr r0
336
337 // XXX Note: we don't really need to call swsusp_resume
338
339 li r3,0
340 blr
341
342/* FIXME:This construct is actually not useful since we don't shut
343 * down the instruction MMU, we could just flip back MSR-DR on.
344 */
345turn_on_mmu:
346 mflr r4
347 mtsrr0 r4
348 mtsrr1 r3
349 sync
350 isync
351 rfi
352