Lines Matching full:s64

1607 static __always_inline s64
1624 static __always_inline s64
1634 * @i: s64 value to assign
1643 atomic64_set(atomic64_t *v, s64 i) in atomic64_set()
1652 * @i: s64 value to assign
1661 atomic64_set_release(atomic64_t *v, s64 i) in atomic64_set_release()
1670 * @i: s64 value to add
1680 atomic64_add(s64 i, atomic64_t *v) in atomic64_add()
1688 * @i: s64 value to add
1697 static __always_inline s64
1698 atomic64_add_return(s64 i, atomic64_t *v) in atomic64_add_return()
1707 * @i: s64 value to add
1716 static __always_inline s64
1717 atomic64_add_return_acquire(s64 i, atomic64_t *v) in atomic64_add_return_acquire()
1725 * @i: s64 value to add
1734 static __always_inline s64
1735 atomic64_add_return_release(s64 i, atomic64_t *v) in atomic64_add_return_release()
1744 * @i: s64 value to add
1753 static __always_inline s64
1754 atomic64_add_return_relaxed(s64 i, atomic64_t *v) in atomic64_add_return_relaxed()
1762 * @i: s64 value to add
1771 static __always_inline s64
1772 atomic64_fetch_add(s64 i, atomic64_t *v) in atomic64_fetch_add()
1781 * @i: s64 value to add
1790 static __always_inline s64
1791 atomic64_fetch_add_acquire(s64 i, atomic64_t *v) in atomic64_fetch_add_acquire()
1799 * @i: s64 value to add
1808 static __always_inline s64
1809 atomic64_fetch_add_release(s64 i, atomic64_t *v) in atomic64_fetch_add_release()
1818 * @i: s64 value to add
1827 static __always_inline s64
1828 atomic64_fetch_add_relaxed(s64 i, atomic64_t *v) in atomic64_fetch_add_relaxed()
1836 * @i: s64 value to subtract
1846 atomic64_sub(s64 i, atomic64_t *v) in atomic64_sub()
1854 * @i: s64 value to subtract
1863 static __always_inline s64
1864 atomic64_sub_return(s64 i, atomic64_t *v) in atomic64_sub_return()
1873 * @i: s64 value to subtract
1882 static __always_inline s64
1883 atomic64_sub_return_acquire(s64 i, atomic64_t *v) in atomic64_sub_return_acquire()
1891 * @i: s64 value to subtract
1900 static __always_inline s64
1901 atomic64_sub_return_release(s64 i, atomic64_t *v) in atomic64_sub_return_release()
1910 * @i: s64 value to subtract
1919 static __always_inline s64
1920 atomic64_sub_return_relaxed(s64 i, atomic64_t *v) in atomic64_sub_return_relaxed()
1928 * @i: s64 value to subtract
1937 static __always_inline s64
1938 atomic64_fetch_sub(s64 i, atomic64_t *v) in atomic64_fetch_sub()
1947 * @i: s64 value to subtract
1956 static __always_inline s64
1957 atomic64_fetch_sub_acquire(s64 i, atomic64_t *v) in atomic64_fetch_sub_acquire()
1965 * @i: s64 value to subtract
1974 static __always_inline s64
1975 atomic64_fetch_sub_release(s64 i, atomic64_t *v) in atomic64_fetch_sub_release()
1984 * @i: s64 value to subtract
1993 static __always_inline s64
1994 atomic64_fetch_sub_relaxed(s64 i, atomic64_t *v) in atomic64_fetch_sub_relaxed()
2027 static __always_inline s64
2045 static __always_inline s64
2062 static __always_inline s64
2080 static __always_inline s64
2097 static __always_inline s64
2115 static __always_inline s64
2132 static __always_inline s64
2150 static __always_inline s64
2184 static __always_inline s64
2202 static __always_inline s64
2219 static __always_inline s64
2237 static __always_inline s64
2254 static __always_inline s64
2272 static __always_inline s64
2289 static __always_inline s64
2307 static __always_inline s64
2316 * @i: s64 value
2326 atomic64_and(s64 i, atomic64_t *v) in atomic64_and()
2334 * @i: s64 value
2343 static __always_inline s64
2344 atomic64_fetch_and(s64 i, atomic64_t *v) in atomic64_fetch_and()
2353 * @i: s64 value
2362 static __always_inline s64
2363 atomic64_fetch_and_acquire(s64 i, atomic64_t *v) in atomic64_fetch_and_acquire()
2371 * @i: s64 value
2380 static __always_inline s64
2381 atomic64_fetch_and_release(s64 i, atomic64_t *v) in atomic64_fetch_and_release()
2390 * @i: s64 value
2399 static __always_inline s64
2400 atomic64_fetch_and_relaxed(s64 i, atomic64_t *v) in atomic64_fetch_and_relaxed()
2408 * @i: s64 value
2418 atomic64_andnot(s64 i, atomic64_t *v) in atomic64_andnot()
2426 * @i: s64 value
2435 static __always_inline s64
2436 atomic64_fetch_andnot(s64 i, atomic64_t *v) in atomic64_fetch_andnot()
2445 * @i: s64 value
2454 static __always_inline s64
2455 atomic64_fetch_andnot_acquire(s64 i, atomic64_t *v) in atomic64_fetch_andnot_acquire()
2463 * @i: s64 value
2472 static __always_inline s64
2473 atomic64_fetch_andnot_release(s64 i, atomic64_t *v) in atomic64_fetch_andnot_release()
2482 * @i: s64 value
2491 static __always_inline s64
2492 atomic64_fetch_andnot_relaxed(s64 i, atomic64_t *v) in atomic64_fetch_andnot_relaxed()
2500 * @i: s64 value
2510 atomic64_or(s64 i, atomic64_t *v) in atomic64_or()
2518 * @i: s64 value
2527 static __always_inline s64
2528 atomic64_fetch_or(s64 i, atomic64_t *v) in atomic64_fetch_or()
2537 * @i: s64 value
2546 static __always_inline s64
2547 atomic64_fetch_or_acquire(s64 i, atomic64_t *v) in atomic64_fetch_or_acquire()
2555 * @i: s64 value
2564 static __always_inline s64
2565 atomic64_fetch_or_release(s64 i, atomic64_t *v) in atomic64_fetch_or_release()
2574 * @i: s64 value
2583 static __always_inline s64
2584 atomic64_fetch_or_relaxed(s64 i, atomic64_t *v) in atomic64_fetch_or_relaxed()
2592 * @i: s64 value
2602 atomic64_xor(s64 i, atomic64_t *v) in atomic64_xor()
2610 * @i: s64 value
2619 static __always_inline s64
2620 atomic64_fetch_xor(s64 i, atomic64_t *v) in atomic64_fetch_xor()
2629 * @i: s64 value
2638 static __always_inline s64
2639 atomic64_fetch_xor_acquire(s64 i, atomic64_t *v) in atomic64_fetch_xor_acquire()
2647 * @i: s64 value
2656 static __always_inline s64
2657 atomic64_fetch_xor_release(s64 i, atomic64_t *v) in atomic64_fetch_xor_release()
2666 * @i: s64 value
2675 static __always_inline s64
2676 atomic64_fetch_xor_relaxed(s64 i, atomic64_t *v) in atomic64_fetch_xor_relaxed()
2685 * @new: s64 value to assign
2693 static __always_inline s64
2694 atomic64_xchg(atomic64_t *v, s64 new) in atomic64_xchg()
2704 * @new: s64 value to assign
2712 static __always_inline s64
2713 atomic64_xchg_acquire(atomic64_t *v, s64 new) in atomic64_xchg_acquire()
2722 * @new: s64 value to assign
2730 static __always_inline s64
2731 atomic64_xchg_release(atomic64_t *v, s64 new) in atomic64_xchg_release()
2741 * @new: s64 value to assign
2749 static __always_inline s64
2750 atomic64_xchg_relaxed(atomic64_t *v, s64 new) in atomic64_xchg_relaxed()
2759 * @old: s64 value to compare with
2760 * @new: s64 value to assign
2769 static __always_inline s64
2770 atomic64_cmpxchg(atomic64_t *v, s64 old, s64 new) in atomic64_cmpxchg()
2780 * @old: s64 value to compare with
2781 * @new: s64 value to assign
2790 static __always_inline s64
2791 atomic64_cmpxchg_acquire(atomic64_t *v, s64 old, s64 new) in atomic64_cmpxchg_acquire()
2800 * @old: s64 value to compare with
2801 * @new: s64 value to assign
2810 static __always_inline s64
2811 atomic64_cmpxchg_release(atomic64_t *v, s64 old, s64 new) in atomic64_cmpxchg_release()
2821 * @old: s64 value to compare with
2822 * @new: s64 value to assign
2831 static __always_inline s64
2832 atomic64_cmpxchg_relaxed(atomic64_t *v, s64 old, s64 new) in atomic64_cmpxchg_relaxed()
2841 * @old: pointer to s64 value to compare with
2842 * @new: s64 value to assign
2853 atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new) in atomic64_try_cmpxchg()
2864 * @old: pointer to s64 value to compare with
2865 * @new: s64 value to assign
2876 atomic64_try_cmpxchg_acquire(atomic64_t *v, s64 *old, s64 new) in atomic64_try_cmpxchg_acquire()
2886 * @old: pointer to s64 value to compare with
2887 * @new: s64 value to assign
2898 atomic64_try_cmpxchg_release(atomic64_t *v, s64 *old, s64 new) in atomic64_try_cmpxchg_release()
2909 * @old: pointer to s64 value to compare with
2910 * @new: s64 value to assign
2921 atomic64_try_cmpxchg_relaxed(atomic64_t *v, s64 *old, s64 new) in atomic64_try_cmpxchg_relaxed()
2930 * @i: s64 value to subtract
2940 atomic64_sub_and_test(s64 i, atomic64_t *v) in atomic64_sub_and_test()
2985 * @i: s64 value to add
2995 atomic64_add_negative(s64 i, atomic64_t *v) in atomic64_add_negative()
3004 * @i: s64 value to add
3014 atomic64_add_negative_acquire(s64 i, atomic64_t *v) in atomic64_add_negative_acquire()
3022 * @i: s64 value to add
3032 atomic64_add_negative_release(s64 i, atomic64_t *v) in atomic64_add_negative_release()
3041 * @i: s64 value to add
3051 atomic64_add_negative_relaxed(s64 i, atomic64_t *v) in atomic64_add_negative_relaxed()
3060 * @a: s64 value to add
3061 * @u: s64 value to compare with
3070 static __always_inline s64
3071 atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u) in atomic64_fetch_add_unless()
3081 * @a: s64 value to add
3082 * @u: s64 value to compare with
3092 atomic64_add_unless(atomic64_t *v, s64 a, s64 u) in atomic64_add_unless()
3167 static __always_inline s64