Lines Matching +full:0 +full:x110

15 	__ulong(map_extra, 0x1ull << 32); /* start of mmap() region */
17 __ulong(map_extra, 0x1ull << 44); /* start of mmap() region */
27 __u32 pid = 0;
30 __u64 __arena_global add64_result = 0;
32 __u32 __arena_global add32_result = 0;
33 __u64 __arena_global add_stack_value_copy = 0;
34 __u64 __arena_global add_stack_result = 0;
41 return 0; in add()
52 return 0; in add()
56 __s64 __arena_global sub64_result = 0;
58 __s32 __arena_global sub32_result = 0;
59 __s64 __arena_global sub_stack_value_copy = 0;
60 __s64 __arena_global sub_stack_result = 0;
67 return 0; in sub()
78 return 0; in sub()
82 _Atomic __u64 __arena_global and64_value = (0x110ull << 32);
83 _Atomic __u32 __arena_global and32_value = 0x110;
85 __u64 __arena_global and64_value = (0x110ull << 32);
86 __u32 __arena_global and32_value = 0x110;
93 return 0; in and()
96 __c11_atomic_fetch_and(&and64_value, 0x011ull << 32, memory_order_relaxed); in and()
97 __c11_atomic_fetch_and(&and32_value, 0x011, memory_order_relaxed); in and()
99 __sync_fetch_and_and(&and64_value, 0x011ull << 32); in and()
100 __sync_fetch_and_and(&and32_value, 0x011); in and()
104 return 0; in and()
108 _Atomic __u32 __arena_global or32_value = 0x110;
109 _Atomic __u64 __arena_global or64_value = (0x110ull << 32);
111 __u32 __arena_global or32_value = 0x110;
112 __u64 __arena_global or64_value = (0x110ull << 32);
119 return 0; in or()
122 __c11_atomic_fetch_or(&or64_value, 0x011ull << 32, memory_order_relaxed); in or()
123 __c11_atomic_fetch_or(&or32_value, 0x011, memory_order_relaxed); in or()
125 __sync_fetch_and_or(&or64_value, 0x011ull << 32); in or()
126 __sync_fetch_and_or(&or32_value, 0x011); in or()
130 return 0; in or()
134 _Atomic __u64 __arena_global xor64_value = (0x110ull << 32);
135 _Atomic __u32 __arena_global xor32_value = 0x110;
137 __u64 __arena_global xor64_value = (0x110ull << 32);
138 __u32 __arena_global xor32_value = 0x110;
145 return 0; in xor()
148 __c11_atomic_fetch_xor(&xor64_value, 0x011ull << 32, memory_order_relaxed); in xor()
149 __c11_atomic_fetch_xor(&xor32_value, 0x011, memory_order_relaxed); in xor()
151 __sync_fetch_and_xor(&xor64_value, 0x011ull << 32); in xor()
152 __sync_fetch_and_xor(&xor32_value, 0x011); in xor()
156 return 0; in xor()
160 __u32 __arena_global cmpxchg32_result_fail = 0;
161 __u32 __arena_global cmpxchg32_result_succeed = 0;
163 __u64 __arena_global cmpxchg64_result_fail = 0;
164 __u64 __arena_global cmpxchg64_result_succeed = 0;
170 return 0; in cmpxchg()
172 cmpxchg64_result_fail = __sync_val_compare_and_swap(&cmpxchg64_value, 0, 3); in cmpxchg()
175 cmpxchg32_result_fail = __sync_val_compare_and_swap(&cmpxchg32_value, 0, 3); in cmpxchg()
179 return 0; in cmpxchg()
183 __u64 __arena_global xchg64_result = 0;
185 __u32 __arena_global xchg32_result = 0;
191 return 0; in xchg()
200 return 0; in xchg()
210 return 0; in uaf()
217 page = bpf_arena_alloc_pages(&arena, NULL, 1, NUMA_NO_NODE, 0); in uaf()
242 uaf_sink += __sync_val_compare_and_swap(page32, 0, 1); in uaf()
268 uaf_sink += __sync_val_compare_and_swap(page64, 0, 1); in uaf()
274 return 0; in uaf()