linux/arch/csky/include/asm/uaccess.h
<<
>>
Prefs
   1/* SPDX-License-Identifier: GPL-2.0 */
   2// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd.
   3
   4#ifndef __ASM_CSKY_UACCESS_H
   5#define __ASM_CSKY_UACCESS_H
   6
   7/*
   8 * User space memory access functions
   9 */
  10#include <linux/compiler.h>
  11#include <linux/errno.h>
  12#include <linux/types.h>
  13#include <linux/sched.h>
  14#include <linux/mm.h>
  15#include <linux/string.h>
  16#include <linux/version.h>
  17#include <asm/segment.h>
  18
  19static inline int access_ok(const void *addr, unsigned long size)
  20{
  21        unsigned long limit = current_thread_info()->addr_limit.seg;
  22
  23        return (((unsigned long)addr < limit) &&
  24                ((unsigned long)(addr + size) < limit));
  25}
  26
  27#define __addr_ok(addr) (access_ok(addr, 0))
  28
  29extern int __put_user_bad(void);
  30
  31/*
  32 * Tell gcc we read from memory instead of writing: this is because
  33 * we do not write to any memory gcc knows about, so there are no
  34 * aliasing issues.
  35 */
  36
  37/*
  38 * These are the main single-value transfer routines.  They automatically
  39 * use the right size if we just have the right pointer type.
  40 *
  41 * This gets kind of ugly. We want to return _two_ values in "get_user()"
  42 * and yet we don't want to do any pointers, because that is too much
  43 * of a performance impact. Thus we have a few rather ugly macros here,
  44 * and hide all the ugliness from the user.
  45 *
  46 * The "__xxx" versions of the user access functions are versions that
  47 * do not verify the address space, that must have been done previously
  48 * with a separate "access_ok()" call (this is used when we do multiple
  49 * accesses to the same area of user memory).
  50 *
  51 * As we use the same address space for kernel and user data on
  52 * Ckcore, we can just do these as direct assignments.  (Of course, the
  53 * exception handling means that it's no longer "just"...)
  54 */
  55
  56#define put_user(x, ptr) \
  57        __put_user_check((x), (ptr), sizeof(*(ptr)))
  58
  59#define __put_user(x, ptr) \
  60        __put_user_nocheck((x), (ptr), sizeof(*(ptr)))
  61
  62#define __ptr(x) ((unsigned long *)(x))
  63
  64#define get_user(x, ptr) \
  65        __get_user_check((x), (ptr), sizeof(*(ptr)))
  66
  67#define __get_user(x, ptr) \
  68        __get_user_nocheck((x), (ptr), sizeof(*(ptr)))
  69
  70#define __put_user_nocheck(x, ptr, size)                                \
  71({                                                                      \
  72        long __pu_err = 0;                                              \
  73        typeof(*(ptr)) *__pu_addr = (ptr);                              \
  74        typeof(*(ptr)) __pu_val = (typeof(*(ptr)))(x);                  \
  75        if (__pu_addr)                                                  \
  76                __put_user_size(__pu_val, (__pu_addr), (size),          \
  77                                __pu_err);                              \
  78        __pu_err;                                                       \
  79})
  80
  81#define __put_user_check(x, ptr, size)                                  \
  82({                                                                      \
  83        long __pu_err = -EFAULT;                                        \
  84        typeof(*(ptr)) *__pu_addr = (ptr);                              \
  85        typeof(*(ptr)) __pu_val = (typeof(*(ptr)))(x);                  \
  86        if (access_ok(__pu_addr, size) && __pu_addr)    \
  87                __put_user_size(__pu_val, __pu_addr, (size), __pu_err); \
  88        __pu_err;                                                       \
  89})
  90
  91#define __put_user_size(x, ptr, size, retval)           \
  92do {                                                    \
  93        retval = 0;                                     \
  94        switch (size) {                                 \
  95        case 1:                                         \
  96                __put_user_asm_b(x, ptr, retval);       \
  97                break;                                  \
  98        case 2:                                         \
  99                __put_user_asm_h(x, ptr, retval);       \
 100                break;                                  \
 101        case 4:                                         \
 102                __put_user_asm_w(x, ptr, retval);       \
 103                break;                                  \
 104        case 8:                                         \
 105                __put_user_asm_64(x, ptr, retval);      \
 106                break;                                  \
 107        default:                                        \
 108                __put_user_bad();                       \
 109        }                                               \
 110} while (0)
 111
 112/*
 113 * We don't tell gcc that we are accessing memory, but this is OK
 114 * because we do not write to any memory gcc knows about, so there
 115 * are no aliasing issues.
 116 *
 117 * Note that PC at a fault is the address *after* the faulting
 118 * instruction.
 119 */
 120#define __put_user_asm_b(x, ptr, err)                   \
 121do {                                                    \
 122        int errcode;                                    \
 123        asm volatile(                                   \
 124        "1:     stb   %1, (%2,0)        \n"             \
 125        "       br    3f                \n"             \
 126        "2:     mov   %0, %3            \n"             \
 127        "       br    3f                \n"             \
 128        ".section __ex_table, \"a\"     \n"             \
 129        ".align   2                     \n"             \
 130        ".long    1b,2b                 \n"             \
 131        ".previous                      \n"             \
 132        "3:                             \n"             \
 133        : "=r"(err), "=r"(x), "=r"(ptr), "=r"(errcode)  \
 134        : "0"(err), "1"(x), "2"(ptr), "3"(-EFAULT)      \
 135        : "memory");                                    \
 136} while (0)
 137
 138#define __put_user_asm_h(x, ptr, err)                   \
 139do {                                                    \
 140        int errcode;                                    \
 141        asm volatile(                                   \
 142        "1:     sth   %1, (%2,0)        \n"             \
 143        "       br    3f                \n"             \
 144        "2:     mov   %0, %3            \n"             \
 145        "       br    3f                \n"             \
 146        ".section __ex_table, \"a\"     \n"             \
 147        ".align   2                     \n"             \
 148        ".long    1b,2b                 \n"             \
 149        ".previous                      \n"             \
 150        "3:                             \n"             \
 151        : "=r"(err), "=r"(x), "=r"(ptr), "=r"(errcode)  \
 152        : "0"(err), "1"(x), "2"(ptr), "3"(-EFAULT)      \
 153        : "memory");                                    \
 154} while (0)
 155
 156#define __put_user_asm_w(x, ptr, err)                   \
 157do {                                                    \
 158        int errcode;                                    \
 159        asm volatile(                                   \
 160        "1:     stw   %1, (%2,0)        \n"             \
 161        "       br    3f                \n"             \
 162        "2:     mov   %0, %3            \n"             \
 163        "       br    3f                \n"             \
 164        ".section __ex_table,\"a\"      \n"             \
 165        ".align   2                     \n"             \
 166        ".long    1b, 2b                \n"             \
 167        ".previous                      \n"             \
 168        "3:                             \n"             \
 169        : "=r"(err), "=r"(x), "=r"(ptr), "=r"(errcode)  \
 170        : "0"(err), "1"(x), "2"(ptr), "3"(-EFAULT)      \
 171        : "memory");                                    \
 172} while (0)
 173
 174#define __put_user_asm_64(x, ptr, err)                          \
 175do {                                                            \
 176        int tmp;                                                \
 177        int errcode;                                            \
 178        typeof(*(ptr))src = (typeof(*(ptr)))x;                  \
 179        typeof(*(ptr))*psrc = &src;                             \
 180                                                                \
 181        asm volatile(                                           \
 182        "     ldw     %3, (%1, 0)     \n"                       \
 183        "1:   stw     %3, (%2, 0)     \n"                       \
 184        "     ldw     %3, (%1, 4)     \n"                       \
 185        "2:   stw     %3, (%2, 4)     \n"                       \
 186        "     br      4f              \n"                       \
 187        "3:   mov     %0, %4          \n"                       \
 188        "     br      4f              \n"                       \
 189        ".section __ex_table, \"a\"   \n"                       \
 190        ".align   2                   \n"                       \
 191        ".long    1b, 3b              \n"                       \
 192        ".long    2b, 3b              \n"                       \
 193        ".previous                    \n"                       \
 194        "4:                           \n"                       \
 195        : "=r"(err), "=r"(psrc), "=r"(ptr),                     \
 196          "=r"(tmp), "=r"(errcode)                              \
 197        : "0"(err), "1"(psrc), "2"(ptr), "3"(0), "4"(-EFAULT)   \
 198        : "memory");                                            \
 199} while (0)
 200
 201#define __get_user_nocheck(x, ptr, size)                        \
 202({                                                              \
 203        long  __gu_err;                                         \
 204        __get_user_size(x, (ptr), (size), __gu_err);            \
 205        __gu_err;                                               \
 206})
 207
 208#define __get_user_check(x, ptr, size)                          \
 209({                                                              \
 210        int __gu_err = -EFAULT;                                 \
 211        const __typeof__(*(ptr)) __user *__gu_ptr = (ptr);      \
 212        if (access_ok(__gu_ptr, size) && __gu_ptr)      \
 213                __get_user_size(x, __gu_ptr, size, __gu_err);   \
 214        __gu_err;                                               \
 215})
 216
 217#define __get_user_size(x, ptr, size, retval)                   \
 218do {                                                            \
 219        switch (size) {                                         \
 220        case 1:                                                 \
 221                __get_user_asm_common((x), ptr, "ldb", retval); \
 222                break;                                          \
 223        case 2:                                                 \
 224                __get_user_asm_common((x), ptr, "ldh", retval); \
 225                break;                                          \
 226        case 4:                                                 \
 227                __get_user_asm_common((x), ptr, "ldw", retval); \
 228                break;                                          \
 229        default:                                                \
 230                x = 0;                                          \
 231                (retval) = __get_user_bad();                    \
 232        }                                                       \
 233} while (0)
 234
 235#define __get_user_asm_common(x, ptr, ins, err)                 \
 236do {                                                            \
 237        int errcode;                                            \
 238        asm volatile(                                           \
 239        "1:   " ins " %1, (%4,0)        \n"                     \
 240        "       br    3f                \n"                     \
 241        /* Fix up codes */                                      \
 242        "2:     mov   %0, %2            \n"                     \
 243        "       movi  %1, 0             \n"                     \
 244        "       br    3f                \n"                     \
 245        ".section __ex_table,\"a\"      \n"                     \
 246        ".align   2                     \n"                     \
 247        ".long    1b, 2b                \n"                     \
 248        ".previous                      \n"                     \
 249        "3:                             \n"                     \
 250        : "=r"(err), "=r"(x), "=r"(errcode)                     \
 251        : "0"(0), "r"(ptr), "2"(-EFAULT)                        \
 252        : "memory");                                            \
 253} while (0)
 254
 255extern int __get_user_bad(void);
 256
 257#define __copy_user(to, from, n)                        \
 258do {                                                    \
 259        int w0, w1, w2, w3;                             \
 260        asm volatile(                                   \
 261        "0:     cmpnei  %1, 0           \n"             \
 262        "       bf      8f              \n"             \
 263        "       mov     %3, %1          \n"             \
 264        "       or      %3, %2          \n"             \
 265        "       andi    %3, 3           \n"             \
 266        "       cmpnei  %3, 0           \n"             \
 267        "       bf      1f              \n"             \
 268        "       br      5f              \n"             \
 269        "1:     cmplti  %0, 16          \n" /* 4W */    \
 270        "       bt      3f              \n"             \
 271        "       ldw     %3, (%2, 0)     \n"             \
 272        "       ldw     %4, (%2, 4)     \n"             \
 273        "       ldw     %5, (%2, 8)     \n"             \
 274        "       ldw     %6, (%2, 12)    \n"             \
 275        "2:     stw     %3, (%1, 0)     \n"             \
 276        "9:     stw     %4, (%1, 4)     \n"             \
 277        "10:    stw     %5, (%1, 8)     \n"             \
 278        "11:    stw     %6, (%1, 12)    \n"             \
 279        "       addi    %2, 16          \n"             \
 280        "       addi    %1, 16          \n"             \
 281        "       subi    %0, 16          \n"             \
 282        "       br      1b              \n"             \
 283        "3:     cmplti  %0, 4           \n" /* 1W */    \
 284        "       bt      5f              \n"             \
 285        "       ldw     %3, (%2, 0)     \n"             \
 286        "4:     stw     %3, (%1, 0)     \n"             \
 287        "       addi    %2, 4           \n"             \
 288        "       addi    %1, 4           \n"             \
 289        "       subi    %0, 4           \n"             \
 290        "       br      3b              \n"             \
 291        "5:     cmpnei  %0, 0           \n"  /* 1B */   \
 292        "       bf      8f              \n"             \
 293        "       ldb     %3, (%2, 0)     \n"             \
 294        "6:     stb     %3, (%1, 0)     \n"             \
 295        "       addi    %2,  1          \n"             \
 296        "       addi    %1,  1          \n"             \
 297        "       subi    %0,  1          \n"             \
 298        "       br      5b              \n"             \
 299        "7:     br      8f              \n"             \
 300        ".section __ex_table, \"a\"     \n"             \
 301        ".align   2                     \n"             \
 302        ".long    2b, 7b                \n"             \
 303        ".long    9b, 7b                \n"             \
 304        ".long   10b, 7b                \n"             \
 305        ".long   11b, 7b                \n"             \
 306        ".long    4b, 7b                \n"             \
 307        ".long    6b, 7b                \n"             \
 308        ".previous                      \n"             \
 309        "8:                             \n"             \
 310        : "=r"(n), "=r"(to), "=r"(from), "=r"(w0),      \
 311          "=r"(w1), "=r"(w2), "=r"(w3)                  \
 312        : "0"(n), "1"(to), "2"(from)                    \
 313        : "memory");                                    \
 314} while (0)
 315
 316#define __copy_user_zeroing(to, from, n)                \
 317do {                                                    \
 318        int tmp;                                        \
 319        int nsave;                                      \
 320        asm volatile(                                   \
 321        "0:     cmpnei  %1, 0           \n"             \
 322        "       bf      7f              \n"             \
 323        "       mov     %3, %1          \n"             \
 324        "       or      %3, %2          \n"             \
 325        "       andi    %3, 3           \n"             \
 326        "       cmpnei  %3, 0           \n"             \
 327        "       bf      1f              \n"             \
 328        "       br      5f              \n"             \
 329        "1:     cmplti  %0, 16          \n"             \
 330        "       bt      3f              \n"             \
 331        "2:     ldw     %3, (%2, 0)     \n"             \
 332        "10:    ldw     %4, (%2, 4)     \n"             \
 333        "       stw     %3, (%1, 0)     \n"             \
 334        "       stw     %4, (%1, 4)     \n"             \
 335        "11:    ldw     %3, (%2, 8)     \n"             \
 336        "12:    ldw     %4, (%2, 12)    \n"             \
 337        "       stw     %3, (%1, 8)     \n"             \
 338        "       stw     %4, (%1, 12)    \n"             \
 339        "       addi    %2, 16          \n"             \
 340        "       addi    %1, 16          \n"             \
 341        "       subi    %0, 16          \n"             \
 342        "       br      1b              \n"             \
 343        "3:     cmplti  %0, 4           \n"             \
 344        "       bt      5f              \n"             \
 345        "4:     ldw     %3, (%2, 0)     \n"             \
 346        "       stw     %3, (%1, 0)     \n"             \
 347        "       addi    %2, 4           \n"             \
 348        "       addi    %1, 4           \n"             \
 349        "       subi    %0, 4           \n"             \
 350        "       br      3b              \n"             \
 351        "5:     cmpnei  %0, 0           \n"             \
 352        "       bf      7f              \n"             \
 353        "6:     ldb     %3, (%2, 0)     \n"             \
 354        "       stb     %3, (%1, 0)     \n"             \
 355        "       addi    %2,  1          \n"             \
 356        "       addi    %1,  1          \n"             \
 357        "       subi    %0,  1          \n"             \
 358        "       br      5b              \n"             \
 359        "8:     mov     %3, %0          \n"             \
 360        "       movi    %4, 0           \n"             \
 361        "9:     stb     %4, (%1, 0)     \n"             \
 362        "       addi    %1, 1           \n"             \
 363        "       subi    %3, 1           \n"             \
 364        "       cmpnei  %3, 0           \n"             \
 365        "       bt      9b              \n"             \
 366        "       br      7f              \n"             \
 367        ".section __ex_table, \"a\"     \n"             \
 368        ".align   2                     \n"             \
 369        ".long    2b, 8b                \n"             \
 370        ".long   10b, 8b                \n"             \
 371        ".long   11b, 8b                \n"             \
 372        ".long   12b, 8b                \n"             \
 373        ".long    4b, 8b                \n"             \
 374        ".long    6b, 8b                \n"             \
 375        ".previous                      \n"             \
 376        "7:                             \n"             \
 377        : "=r"(n), "=r"(to), "=r"(from), "=r"(nsave),   \
 378          "=r"(tmp)                                     \
 379        : "0"(n), "1"(to), "2"(from)                    \
 380        : "memory");                                    \
 381} while (0)
 382
 383unsigned long raw_copy_from_user(void *to, const void *from, unsigned long n);
 384unsigned long raw_copy_to_user(void *to, const void *from, unsigned long n);
 385
 386unsigned long clear_user(void *to, unsigned long n);
 387unsigned long __clear_user(void __user *to, unsigned long n);
 388
 389long strncpy_from_user(char *dst, const char *src, long count);
 390long __strncpy_from_user(char *dst, const char *src, long count);
 391
 392/*
 393 * Return the size of a string (including the ending 0)
 394 *
 395 * Return 0 on exception, a value greater than N if too long
 396 */
 397long strnlen_user(const char *src, long n);
 398
 399#define strlen_user(str) strnlen_user(str, 32767)
 400
 401struct exception_table_entry {
 402        unsigned long insn;
 403        unsigned long nextinsn;
 404};
 405
 406extern int fixup_exception(struct pt_regs *regs);
 407
 408#endif /* __ASM_CSKY_UACCESS_H */
 409