|
5 | 5 | #ifndef __ASM_BARRIER_H
|
6 | 6 | #define __ASM_BARRIER_H
|
7 | 7 |
|
8 |
| -#define __sync() __asm__ __volatile__("dbar 0" : : : "memory") |
| 8 | +/* |
| 9 | + * Hint encoding: |
| 10 | + * |
| 11 | + * Bit4: ordering or completion (0: completion, 1: ordering) |
| 12 | + * Bit3: barrier for previous read (0: true, 1: false) |
| 13 | + * Bit2: barrier for previous write (0: true, 1: false) |
| 14 | + * Bit1: barrier for succeeding read (0: true, 1: false) |
| 15 | + * Bit0: barrier for succeeding write (0: true, 1: false) |
| 16 | + * |
| 17 | + * Hint 0x700: barrier for "read after read" from the same address |
| 18 | + */ |
| 19 | + |
| 20 | +#define DBAR(hint) __asm__ __volatile__("dbar %0 " : : "I"(hint) : "memory") |
| 21 | + |
| 22 | +#define crwrw 0b00000 |
| 23 | +#define cr_r_ 0b00101 |
| 24 | +#define c_w_w 0b01010 |
9 | 25 |
|
10 |
| -#define fast_wmb() __sync() |
11 |
| -#define fast_rmb() __sync() |
12 |
| -#define fast_mb() __sync() |
13 |
| -#define fast_iob() __sync() |
14 |
| -#define wbflush() __sync() |
| 26 | +#define orwrw 0b10000 |
| 27 | +#define or_r_ 0b10101 |
| 28 | +#define o_w_w 0b11010 |
15 | 29 |
|
16 |
| -#define wmb() fast_wmb() |
17 |
| -#define rmb() fast_rmb() |
18 |
| -#define mb() fast_mb() |
19 |
| -#define iob() fast_iob() |
| 30 | +#define orw_w 0b10010 |
| 31 | +#define or_rw 0b10100 |
20 | 32 |
|
21 |
| -#define __smp_mb() __asm__ __volatile__("dbar 0" : : : "memory") |
22 |
| -#define __smp_rmb() __asm__ __volatile__("dbar 0" : : : "memory") |
23 |
| -#define __smp_wmb() __asm__ __volatile__("dbar 0" : : : "memory") |
| 33 | +#define c_sync() DBAR(crwrw) |
| 34 | +#define c_rsync() DBAR(cr_r_) |
| 35 | +#define c_wsync() DBAR(c_w_w) |
| 36 | + |
| 37 | +#define o_sync() DBAR(orwrw) |
| 38 | +#define o_rsync() DBAR(or_r_) |
| 39 | +#define o_wsync() DBAR(o_w_w) |
| 40 | + |
| 41 | +#define ldacq_mb() DBAR(or_rw) |
| 42 | +#define strel_mb() DBAR(orw_w) |
| 43 | + |
| 44 | +#define mb() c_sync() |
| 45 | +#define rmb() c_rsync() |
| 46 | +#define wmb() c_wsync() |
| 47 | +#define iob() c_sync() |
| 48 | +#define wbflush() c_sync() |
| 49 | + |
| 50 | +#define __smp_mb() o_sync() |
| 51 | +#define __smp_rmb() o_rsync() |
| 52 | +#define __smp_wmb() o_wsync() |
24 | 53 |
|
25 | 54 | #ifdef CONFIG_SMP
|
26 |
| -#define __WEAK_LLSC_MB " dbar 0 \n" |
| 55 | +#define __WEAK_LLSC_MB " dbar 0x700 \n" |
27 | 56 | #else
|
28 |
| -#define __WEAK_LLSC_MB " \n" |
| 57 | +#define __WEAK_LLSC_MB " \n" |
29 | 58 | #endif
|
30 | 59 |
|
31 | 60 | #define __smp_mb__before_atomic() barrier()
|
@@ -59,68 +88,19 @@ static inline unsigned long array_index_mask_nospec(unsigned long index,
|
59 | 88 | return mask;
|
60 | 89 | }
|
61 | 90 |
|
62 |
| -#define __smp_load_acquire(p) \ |
63 |
| -({ \ |
64 |
| - union { typeof(*p) __val; char __c[1]; } __u; \ |
65 |
| - unsigned long __tmp = 0; \ |
66 |
| - compiletime_assert_atomic_type(*p); \ |
67 |
| - switch (sizeof(*p)) { \ |
68 |
| - case 1: \ |
69 |
| - *(__u8 *)__u.__c = *(volatile __u8 *)p; \ |
70 |
| - __smp_mb(); \ |
71 |
| - break; \ |
72 |
| - case 2: \ |
73 |
| - *(__u16 *)__u.__c = *(volatile __u16 *)p; \ |
74 |
| - __smp_mb(); \ |
75 |
| - break; \ |
76 |
| - case 4: \ |
77 |
| - __asm__ __volatile__( \ |
78 |
| - "amor_db.w %[val], %[tmp], %[mem] \n" \ |
79 |
| - : [val] "=&r" (*(__u32 *)__u.__c) \ |
80 |
| - : [mem] "ZB" (*(u32 *) p), [tmp] "r" (__tmp) \ |
81 |
| - : "memory"); \ |
82 |
| - break; \ |
83 |
| - case 8: \ |
84 |
| - __asm__ __volatile__( \ |
85 |
| - "amor_db.d %[val], %[tmp], %[mem] \n" \ |
86 |
| - : [val] "=&r" (*(__u64 *)__u.__c) \ |
87 |
| - : [mem] "ZB" (*(u64 *) p), [tmp] "r" (__tmp) \ |
88 |
| - : "memory"); \ |
89 |
| - break; \ |
90 |
| - } \ |
91 |
| - (typeof(*p))__u.__val; \ |
| 91 | +#define __smp_load_acquire(p) \ |
| 92 | +({ \ |
| 93 | + typeof(*p) ___p1 = READ_ONCE(*p); \ |
| 94 | + compiletime_assert_atomic_type(*p); \ |
| 95 | + ldacq_mb(); \ |
| 96 | + ___p1; \ |
92 | 97 | })
|
93 | 98 |
|
94 |
| -#define __smp_store_release(p, v) \ |
95 |
| -do { \ |
96 |
| - union { typeof(*p) __val; char __c[1]; } __u = \ |
97 |
| - { .__val = (__force typeof(*p)) (v) }; \ |
98 |
| - unsigned long __tmp; \ |
99 |
| - compiletime_assert_atomic_type(*p); \ |
100 |
| - switch (sizeof(*p)) { \ |
101 |
| - case 1: \ |
102 |
| - __smp_mb(); \ |
103 |
| - *(volatile __u8 *)p = *(__u8 *)__u.__c; \ |
104 |
| - break; \ |
105 |
| - case 2: \ |
106 |
| - __smp_mb(); \ |
107 |
| - *(volatile __u16 *)p = *(__u16 *)__u.__c; \ |
108 |
| - break; \ |
109 |
| - case 4: \ |
110 |
| - __asm__ __volatile__( \ |
111 |
| - "amswap_db.w %[tmp], %[val], %[mem] \n" \ |
112 |
| - : [mem] "+ZB" (*(u32 *)p), [tmp] "=&r" (__tmp) \ |
113 |
| - : [val] "r" (*(__u32 *)__u.__c) \ |
114 |
| - : ); \ |
115 |
| - break; \ |
116 |
| - case 8: \ |
117 |
| - __asm__ __volatile__( \ |
118 |
| - "amswap_db.d %[tmp], %[val], %[mem] \n" \ |
119 |
| - : [mem] "+ZB" (*(u64 *)p), [tmp] "=&r" (__tmp) \ |
120 |
| - : [val] "r" (*(__u64 *)__u.__c) \ |
121 |
| - : ); \ |
122 |
| - break; \ |
123 |
| - } \ |
| 99 | +#define __smp_store_release(p, v) \ |
| 100 | +do { \ |
| 101 | + compiletime_assert_atomic_type(*p); \ |
| 102 | + strel_mb(); \ |
| 103 | + WRITE_ONCE(*p, v); \ |
124 | 104 | } while (0)
|
125 | 105 |
|
126 | 106 | #define __smp_store_mb(p, v) \
|
|
0 commit comments