Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | #ifndef __M68K_UACCESS_H |
2 | #define __M68K_UACCESS_H | |
3 | ||
4 | /* | |
5 | * User space memory access functions | |
6 | */ | |
d94af931 | 7 | #include <linux/compiler.h> |
1da177e4 | 8 | #include <linux/errno.h> |
d94af931 | 9 | #include <linux/types.h> |
a0f7b67a | 10 | #include <linux/sched.h> |
1da177e4 LT |
11 | #include <asm/segment.h> |
12 | ||
13 | #define VERIFY_READ 0 | |
14 | #define VERIFY_WRITE 1 | |
15 | ||
16 | /* We let the MMU do all checking */ | |
70f9cac5 GU |
17 | static inline int access_ok(int type, const void __user *addr, |
18 | unsigned long size) | |
19 | { | |
20 | return 1; | |
21 | } | |
1da177e4 | 22 | |
e08d703c GU |
23 | /* |
24 | * Not all varients of the 68k family support the notion of address spaces. | |
25 | * The traditional 680x0 parts do, and they use the sfc/dfc registers and | |
26 | * the "moves" instruction to access user space from kernel space. Other | |
27 | * family members like ColdFire don't support this, and only have a single | |
28 | * address space, and use the usual "move" instruction for user space access. | |
29 | * | |
30 | * Outside of this difference the user space access functions are the same. | |
31 | * So lets keep the code simple and just define in what we need to use. | |
32 | */ | |
33 | #ifdef CONFIG_CPU_HAS_ADDRESS_SPACES | |
34 | #define MOVES "moves" | |
35 | #else | |
36 | #define MOVES "move" | |
37 | #endif | |
38 | ||
1da177e4 LT |
39 | /* |
40 | * The exception table consists of pairs of addresses: the first is the | |
41 | * address of an instruction that is allowed to fault, and the second is | |
42 | * the address at which the program should continue. No registers are | |
43 | * modified, so it is entirely up to the continuation code to figure out | |
44 | * what to do. | |
45 | * | |
46 | * All the routines below use bits of fixup code that are out of line | |
47 | * with the main instruction path. This means when everything is well, | |
48 | * we don't even have to jump over them. Further, they do not intrude | |
49 | * on our cache or tlb entries. | |
50 | */ | |
51 | ||
52 | struct exception_table_entry | |
53 | { | |
54 | unsigned long insn, fixup; | |
55 | }; | |
56 | ||
d94af931 RZ |
57 | extern int __put_user_bad(void); |
58 | extern int __get_user_bad(void); | |
59 | ||
60 | #define __put_user_asm(res, x, ptr, bwl, reg, err) \ | |
61 | asm volatile ("\n" \ | |
e08d703c | 62 | "1: "MOVES"."#bwl" %2,%1\n" \ |
d94af931 RZ |
63 | "2:\n" \ |
64 | " .section .fixup,\"ax\"\n" \ | |
65 | " .even\n" \ | |
66 | "10: moveq.l %3,%0\n" \ | |
67 | " jra 2b\n" \ | |
68 | " .previous\n" \ | |
69 | "\n" \ | |
70 | " .section __ex_table,\"a\"\n" \ | |
71 | " .align 4\n" \ | |
72 | " .long 1b,10b\n" \ | |
73 | " .long 2b,10b\n" \ | |
74 | " .previous" \ | |
75 | : "+d" (res), "=m" (*(ptr)) \ | |
76 | : #reg (x), "i" (err)) | |
1da177e4 LT |
77 | |
78 | /* | |
79 | * These are the main single-value transfer routines. They automatically | |
80 | * use the right size if we just have the right pointer type. | |
81 | */ | |
82 | ||
d94af931 RZ |
83 | #define __put_user(x, ptr) \ |
84 | ({ \ | |
85 | typeof(*(ptr)) __pu_val = (x); \ | |
86 | int __pu_err = 0; \ | |
87 | __chk_user_ptr(ptr); \ | |
88 | switch (sizeof (*(ptr))) { \ | |
89 | case 1: \ | |
90 | __put_user_asm(__pu_err, __pu_val, ptr, b, d, -EFAULT); \ | |
91 | break; \ | |
92 | case 2: \ | |
631d8b67 | 93 | __put_user_asm(__pu_err, __pu_val, ptr, w, r, -EFAULT); \ |
d94af931 RZ |
94 | break; \ |
95 | case 4: \ | |
96 | __put_user_asm(__pu_err, __pu_val, ptr, l, r, -EFAULT); \ | |
97 | break; \ | |
98 | case 8: \ | |
99 | { \ | |
b971018b | 100 | const void __user *__pu_ptr = (ptr); \ |
d94af931 | 101 | asm volatile ("\n" \ |
e08d703c GU |
102 | "1: "MOVES".l %2,(%1)+\n" \ |
103 | "2: "MOVES".l %R2,(%1)\n" \ | |
d94af931 RZ |
104 | "3:\n" \ |
105 | " .section .fixup,\"ax\"\n" \ | |
106 | " .even\n" \ | |
107 | "10: movel %3,%0\n" \ | |
108 | " jra 3b\n" \ | |
109 | " .previous\n" \ | |
110 | "\n" \ | |
111 | " .section __ex_table,\"a\"\n" \ | |
112 | " .align 4\n" \ | |
113 | " .long 1b,10b\n" \ | |
114 | " .long 2b,10b\n" \ | |
115 | " .long 3b,10b\n" \ | |
116 | " .previous" \ | |
117 | : "+d" (__pu_err), "+a" (__pu_ptr) \ | |
118 | : "r" (__pu_val), "i" (-EFAULT) \ | |
119 | : "memory"); \ | |
120 | break; \ | |
121 | } \ | |
122 | default: \ | |
123 | __pu_err = __put_user_bad(); \ | |
124 | break; \ | |
125 | } \ | |
126 | __pu_err; \ | |
1da177e4 | 127 | }) |
d94af931 | 128 | #define put_user(x, ptr) __put_user(x, ptr) |
1da177e4 | 129 | |
1da177e4 | 130 | |
d94af931 RZ |
131 | #define __get_user_asm(res, x, ptr, type, bwl, reg, err) ({ \ |
132 | type __gu_val; \ | |
133 | asm volatile ("\n" \ | |
e08d703c | 134 | "1: "MOVES"."#bwl" %2,%1\n" \ |
d94af931 RZ |
135 | "2:\n" \ |
136 | " .section .fixup,\"ax\"\n" \ | |
137 | " .even\n" \ | |
138 | "10: move.l %3,%0\n" \ | |
e08d703c | 139 | " sub.l %1,%1\n" \ |
d94af931 RZ |
140 | " jra 2b\n" \ |
141 | " .previous\n" \ | |
142 | "\n" \ | |
143 | " .section __ex_table,\"a\"\n" \ | |
144 | " .align 4\n" \ | |
145 | " .long 1b,10b\n" \ | |
146 | " .previous" \ | |
147 | : "+d" (res), "=&" #reg (__gu_val) \ | |
148 | : "m" (*(ptr)), "i" (err)); \ | |
b971018b | 149 | (x) = (typeof(*(ptr)))(unsigned long)__gu_val; \ |
1da177e4 | 150 | }) |
1da177e4 | 151 | |
d94af931 RZ |
152 | #define __get_user(x, ptr) \ |
153 | ({ \ | |
154 | int __gu_err = 0; \ | |
155 | __chk_user_ptr(ptr); \ | |
156 | switch (sizeof(*(ptr))) { \ | |
157 | case 1: \ | |
158 | __get_user_asm(__gu_err, x, ptr, u8, b, d, -EFAULT); \ | |
159 | break; \ | |
160 | case 2: \ | |
631d8b67 | 161 | __get_user_asm(__gu_err, x, ptr, u16, w, r, -EFAULT); \ |
d94af931 RZ |
162 | break; \ |
163 | case 4: \ | |
164 | __get_user_asm(__gu_err, x, ptr, u32, l, r, -EFAULT); \ | |
165 | break; \ | |
166 | /* case 8: disabled because gcc-4.1 has a broken typeof \ | |
167 | { \ | |
168 | const void *__gu_ptr = (ptr); \ | |
169 | u64 __gu_val; \ | |
170 | asm volatile ("\n" \ | |
e08d703c GU |
171 | "1: "MOVES".l (%2)+,%1\n" \ |
172 | "2: "MOVES".l (%2),%R1\n" \ | |
d94af931 RZ |
173 | "3:\n" \ |
174 | " .section .fixup,\"ax\"\n" \ | |
175 | " .even\n" \ | |
176 | "10: move.l %3,%0\n" \ | |
177 | " sub.l %1,%1\n" \ | |
178 | " sub.l %R1,%R1\n" \ | |
179 | " jra 3b\n" \ | |
180 | " .previous\n" \ | |
181 | "\n" \ | |
182 | " .section __ex_table,\"a\"\n" \ | |
183 | " .align 4\n" \ | |
184 | " .long 1b,10b\n" \ | |
185 | " .long 2b,10b\n" \ | |
186 | " .previous" \ | |
187 | : "+d" (__gu_err), "=&r" (__gu_val), \ | |
188 | "+a" (__gu_ptr) \ | |
189 | : "i" (-EFAULT) \ | |
190 | : "memory"); \ | |
191 | (x) = (typeof(*(ptr)))__gu_val; \ | |
192 | break; \ | |
193 | } */ \ | |
194 | default: \ | |
195 | __gu_err = __get_user_bad(); \ | |
196 | break; \ | |
197 | } \ | |
198 | __gu_err; \ | |
199 | }) | |
200 | #define get_user(x, ptr) __get_user(x, ptr) | |
1da177e4 | 201 | |
d94af931 RZ |
202 | unsigned long __generic_copy_from_user(void *to, const void __user *from, unsigned long n); |
203 | unsigned long __generic_copy_to_user(void __user *to, const void *from, unsigned long n); | |
1da177e4 | 204 | |
53617825 RZ |
205 | #define __constant_copy_from_user_asm(res, to, from, tmp, n, s1, s2, s3)\ |
206 | asm volatile ("\n" \ | |
e08d703c | 207 | "1: "MOVES"."#s1" (%2)+,%3\n" \ |
53617825 | 208 | " move."#s1" %3,(%1)+\n" \ |
e08d703c | 209 | "2: "MOVES"."#s2" (%2)+,%3\n" \ |
53617825 RZ |
210 | " move."#s2" %3,(%1)+\n" \ |
211 | " .ifnc \""#s3"\",\"\"\n" \ | |
e08d703c | 212 | "3: "MOVES"."#s3" (%2)+,%3\n" \ |
53617825 RZ |
213 | " move."#s3" %3,(%1)+\n" \ |
214 | " .endif\n" \ | |
215 | "4:\n" \ | |
216 | " .section __ex_table,\"a\"\n" \ | |
217 | " .align 4\n" \ | |
218 | " .long 1b,10f\n" \ | |
219 | " .long 2b,20f\n" \ | |
220 | " .ifnc \""#s3"\",\"\"\n" \ | |
221 | " .long 3b,30f\n" \ | |
222 | " .endif\n" \ | |
223 | " .previous\n" \ | |
224 | "\n" \ | |
225 | " .section .fixup,\"ax\"\n" \ | |
226 | " .even\n" \ | |
227 | "10: clr."#s1" (%1)+\n" \ | |
228 | "20: clr."#s2" (%1)+\n" \ | |
229 | " .ifnc \""#s3"\",\"\"\n" \ | |
230 | "30: clr."#s3" (%1)+\n" \ | |
231 | " .endif\n" \ | |
232 | " moveq.l #"#n",%0\n" \ | |
233 | " jra 4b\n" \ | |
234 | " .previous\n" \ | |
235 | : "+d" (res), "+&a" (to), "+a" (from), "=&d" (tmp) \ | |
236 | : : "memory") | |
237 | ||
d94af931 RZ |
238 | static __always_inline unsigned long |
239 | __constant_copy_from_user(void *to, const void __user *from, unsigned long n) | |
1da177e4 | 240 | { |
d94af931 | 241 | unsigned long res = 0, tmp; |
1da177e4 | 242 | |
d94af931 | 243 | switch (n) { |
1da177e4 | 244 | case 1: |
b971018b | 245 | __get_user_asm(res, *(u8 *)to, (u8 __user *)from, u8, b, d, 1); |
53617825 | 246 | break; |
1da177e4 | 247 | case 2: |
631d8b67 | 248 | __get_user_asm(res, *(u16 *)to, (u16 __user *)from, u16, w, r, 2); |
53617825 RZ |
249 | break; |
250 | case 3: | |
251 | __constant_copy_from_user_asm(res, to, from, tmp, 3, w, b,); | |
252 | break; | |
d94af931 | 253 | case 4: |
b971018b | 254 | __get_user_asm(res, *(u32 *)to, (u32 __user *)from, u32, l, r, 4); |
53617825 RZ |
255 | break; |
256 | case 5: | |
257 | __constant_copy_from_user_asm(res, to, from, tmp, 5, l, b,); | |
258 | break; | |
259 | case 6: | |
260 | __constant_copy_from_user_asm(res, to, from, tmp, 6, l, w,); | |
261 | break; | |
262 | case 7: | |
263 | __constant_copy_from_user_asm(res, to, from, tmp, 7, l, w, b); | |
264 | break; | |
265 | case 8: | |
266 | __constant_copy_from_user_asm(res, to, from, tmp, 8, l, l,); | |
267 | break; | |
268 | case 9: | |
269 | __constant_copy_from_user_asm(res, to, from, tmp, 9, l, l, b); | |
270 | break; | |
271 | case 10: | |
272 | __constant_copy_from_user_asm(res, to, from, tmp, 10, l, l, w); | |
273 | break; | |
274 | case 12: | |
275 | __constant_copy_from_user_asm(res, to, from, tmp, 12, l, l, l); | |
276 | break; | |
277 | default: | |
278 | /* we limit the inlined version to 3 moves */ | |
279 | return __generic_copy_from_user(to, from, n); | |
1da177e4 | 280 | } |
1da177e4 | 281 | |
d94af931 RZ |
282 | return res; |
283 | } | |
1da177e4 | 284 | |
53617825 RZ |
285 | #define __constant_copy_to_user_asm(res, to, from, tmp, n, s1, s2, s3) \ |
286 | asm volatile ("\n" \ | |
287 | " move."#s1" (%2)+,%3\n" \ | |
e08d703c | 288 | "11: "MOVES"."#s1" %3,(%1)+\n" \ |
53617825 | 289 | "12: move."#s2" (%2)+,%3\n" \ |
e08d703c | 290 | "21: "MOVES"."#s2" %3,(%1)+\n" \ |
53617825 RZ |
291 | "22:\n" \ |
292 | " .ifnc \""#s3"\",\"\"\n" \ | |
293 | " move."#s3" (%2)+,%3\n" \ | |
e08d703c | 294 | "31: "MOVES"."#s3" %3,(%1)+\n" \ |
53617825 RZ |
295 | "32:\n" \ |
296 | " .endif\n" \ | |
297 | "4:\n" \ | |
298 | "\n" \ | |
299 | " .section __ex_table,\"a\"\n" \ | |
300 | " .align 4\n" \ | |
301 | " .long 11b,5f\n" \ | |
302 | " .long 12b,5f\n" \ | |
303 | " .long 21b,5f\n" \ | |
304 | " .long 22b,5f\n" \ | |
305 | " .ifnc \""#s3"\",\"\"\n" \ | |
306 | " .long 31b,5f\n" \ | |
307 | " .long 32b,5f\n" \ | |
308 | " .endif\n" \ | |
309 | " .previous\n" \ | |
310 | "\n" \ | |
311 | " .section .fixup,\"ax\"\n" \ | |
312 | " .even\n" \ | |
313 | "5: moveq.l #"#n",%0\n" \ | |
314 | " jra 4b\n" \ | |
315 | " .previous\n" \ | |
316 | : "+d" (res), "+a" (to), "+a" (from), "=&d" (tmp) \ | |
317 | : : "memory") | |
318 | ||
d94af931 | 319 | static __always_inline unsigned long |
11c40f8a | 320 | __constant_copy_to_user(void __user *to, const void *from, unsigned long n) |
1da177e4 | 321 | { |
d94af931 RZ |
322 | unsigned long res = 0, tmp; |
323 | ||
d94af931 | 324 | switch (n) { |
1da177e4 | 325 | case 1: |
b971018b | 326 | __put_user_asm(res, *(u8 *)from, (u8 __user *)to, b, d, 1); |
53617825 | 327 | break; |
1da177e4 | 328 | case 2: |
631d8b67 | 329 | __put_user_asm(res, *(u16 *)from, (u16 __user *)to, w, r, 2); |
53617825 RZ |
330 | break; |
331 | case 3: | |
332 | __constant_copy_to_user_asm(res, to, from, tmp, 3, w, b,); | |
333 | break; | |
d94af931 | 334 | case 4: |
b971018b | 335 | __put_user_asm(res, *(u32 *)from, (u32 __user *)to, l, r, 4); |
53617825 RZ |
336 | break; |
337 | case 5: | |
338 | __constant_copy_to_user_asm(res, to, from, tmp, 5, l, b,); | |
339 | break; | |
340 | case 6: | |
341 | __constant_copy_to_user_asm(res, to, from, tmp, 6, l, w,); | |
342 | break; | |
343 | case 7: | |
344 | __constant_copy_to_user_asm(res, to, from, tmp, 7, l, w, b); | |
345 | break; | |
346 | case 8: | |
347 | __constant_copy_to_user_asm(res, to, from, tmp, 8, l, l,); | |
348 | break; | |
349 | case 9: | |
350 | __constant_copy_to_user_asm(res, to, from, tmp, 9, l, l, b); | |
351 | break; | |
352 | case 10: | |
353 | __constant_copy_to_user_asm(res, to, from, tmp, 10, l, l, w); | |
354 | break; | |
355 | case 12: | |
356 | __constant_copy_to_user_asm(res, to, from, tmp, 12, l, l, l); | |
357 | break; | |
358 | default: | |
359 | /* limit the inlined version to 3 moves */ | |
360 | return __generic_copy_to_user(to, from, n); | |
1da177e4 | 361 | } |
d94af931 | 362 | |
d94af931 | 363 | return res; |
1da177e4 LT |
364 | } |
365 | ||
d94af931 | 366 | #define __copy_from_user(to, from, n) \ |
1da177e4 LT |
367 | (__builtin_constant_p(n) ? \ |
368 | __constant_copy_from_user(to, from, n) : \ | |
369 | __generic_copy_from_user(to, from, n)) | |
370 | ||
d94af931 | 371 | #define __copy_to_user(to, from, n) \ |
1da177e4 LT |
372 | (__builtin_constant_p(n) ? \ |
373 | __constant_copy_to_user(to, from, n) : \ | |
374 | __generic_copy_to_user(to, from, n)) | |
375 | ||
d94af931 RZ |
376 | #define __copy_to_user_inatomic __copy_to_user |
377 | #define __copy_from_user_inatomic __copy_from_user | |
1da177e4 | 378 | |
d94af931 RZ |
379 | #define copy_from_user(to, from, n) __copy_from_user(to, from, n) |
380 | #define copy_to_user(to, from, n) __copy_to_user(to, from, n) | |
1da177e4 | 381 | |
d8ce7263 GU |
382 | #define user_addr_max() \ |
383 | (segment_eq(get_fs(), USER_DS) ? TASK_SIZE : ~0UL) | |
384 | ||
385 | extern long strncpy_from_user(char *dst, const char __user *src, long count); | |
386 | extern __must_check long strlen_user(const char __user *str); | |
387 | extern __must_check long strnlen_user(const char __user *str, long n); | |
388 | ||
3c46bdca GU |
389 | unsigned long __clear_user(void __user *to, unsigned long n); |
390 | ||
391 | #define clear_user __clear_user | |
1da177e4 | 392 | |
1da177e4 | 393 | #endif /* _M68K_UACCESS_H */ |