Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 | #ifndef __ASM_SH_ATOMIC_GRB_H #define __ASM_SH_ATOMIC_GRB_H static inline void atomic_add(int i, atomic_t *v) { int tmp; __asm__ __volatile__ ( " .align 2 \n\t" " mova 1f, r0 \n\t" /* r0 = end point */ " mov r15, r1 \n\t" /* r1 = saved sp */ " mov #-6, r15 \n\t" /* LOGIN: r15 = size */ " mov.l @%1, %0 \n\t" /* load old value */ " add %2, %0 \n\t" /* add */ " mov.l %0, @%1 \n\t" /* store new value */ "1: mov r1, r15 \n\t" /* LOGOUT */ : "=&r" (tmp), "+r" (v) : "r" (i) : "memory" , "r0", "r1"); } static inline void atomic_sub(int i, atomic_t *v) { int tmp; __asm__ __volatile__ ( " .align 2 \n\t" " mova 1f, r0 \n\t" /* r0 = end point */ " mov r15, r1 \n\t" /* r1 = saved sp */ " mov #-6, r15 \n\t" /* LOGIN: r15 = size */ " mov.l @%1, %0 \n\t" /* load old value */ " sub %2, %0 \n\t" /* sub */ " mov.l %0, @%1 \n\t" /* store new value */ "1: mov r1, r15 \n\t" /* LOGOUT */ : "=&r" (tmp), "+r" (v) : "r" (i) : "memory" , "r0", "r1"); } static inline int atomic_add_return(int i, atomic_t *v) { int tmp; __asm__ __volatile__ ( " .align 2 \n\t" " mova 1f, r0 \n\t" /* r0 = end point */ " mov r15, r1 \n\t" /* r1 = saved sp */ " mov #-6, r15 \n\t" /* LOGIN: r15 = size */ " mov.l @%1, %0 \n\t" /* load old value */ " add %2, %0 \n\t" /* add */ " mov.l %0, @%1 \n\t" /* store new value */ "1: mov r1, r15 \n\t" /* LOGOUT */ : "=&r" (tmp), "+r" (v) : "r" (i) : "memory" , "r0", "r1"); return tmp; } static inline int atomic_sub_return(int i, atomic_t *v) { int tmp; __asm__ __volatile__ ( " .align 2 \n\t" " mova 1f, r0 \n\t" /* r0 = end point */ " mov r15, r1 \n\t" /* r1 = saved sp */ " mov #-6, r15 \n\t" /* LOGIN: r15 = size */ " mov.l @%1, %0 \n\t" /* load old value */ " sub %2, %0 \n\t" /* sub */ " mov.l %0, @%1 \n\t" /* store new value */ "1: mov r1, r15 \n\t" /* LOGOUT */ : "=&r" (tmp), "+r" (v) : "r" (i) : "memory", "r0", "r1"); return tmp; } static inline void atomic_clear_mask(unsigned int mask, atomic_t *v) { int tmp; unsigned int _mask = ~mask; __asm__ __volatile__ ( " .align 2 \n\t" " mova 1f, r0 \n\t" /* r0 = end point */ " mov r15, r1 \n\t" /* r1 = saved sp */ " mov #-6, r15 \n\t" /* LOGIN: r15 = size */ " mov.l @%1, %0 \n\t" /* load old value */ " and %2, %0 \n\t" /* add */ " mov.l %0, @%1 \n\t" /* store new value */ "1: mov r1, r15 \n\t" /* LOGOUT */ : "=&r" (tmp), "+r" (v) : "r" (_mask) : "memory" , "r0", "r1"); } static inline void atomic_set_mask(unsigned int mask, atomic_t *v) { int tmp; __asm__ __volatile__ ( " .align 2 \n\t" " mova 1f, r0 \n\t" /* r0 = end point */ " mov r15, r1 \n\t" /* r1 = saved sp */ " mov #-6, r15 \n\t" /* LOGIN: r15 = size */ " mov.l @%1, %0 \n\t" /* load old value */ " or %2, %0 \n\t" /* or */ " mov.l %0, @%1 \n\t" /* store new value */ "1: mov r1, r15 \n\t" /* LOGOUT */ : "=&r" (tmp), "+r" (v) : "r" (mask) : "memory" , "r0", "r1"); } #endif /* __ASM_SH_ATOMIC_GRB_H */ |