diff options
author | rth <rth@138bc75d-0d04-0410-961f-82ee72b054a4> | 2012-05-01 15:48:28 +0000 |
---|---|---|
committer | rth <rth@138bc75d-0d04-0410-961f-82ee72b054a4> | 2012-05-01 15:48:28 +0000 |
commit | 6db1d2ca83f9327020b75f6c1610649bb18e71db (patch) | |
tree | 729dc91fa97f38042aa7777ca6267ebed2edb3e8 /libatomic/gexch.c | |
parent | d8219307493056ea48203be157f0cfb468698f2d (diff) | |
download | gcc-6db1d2ca83f9327020b75f6c1610649bb18e71db.tar.gz |
Add libatomic as a target library.
git-svn-id: svn+ssh://gcc.gnu.org/svn/gcc/trunk@187018 138bc75d-0d04-0410-961f-82ee72b054a4
Diffstat (limited to 'libatomic/gexch.c')
-rw-r--r-- | libatomic/gexch.c | 144 |
1 files changed, 144 insertions, 0 deletions
diff --git a/libatomic/gexch.c b/libatomic/gexch.c new file mode 100644 index 00000000000..0125b585714 --- /dev/null +++ b/libatomic/gexch.c @@ -0,0 +1,144 @@ +/* Copyright (C) 2012 Free Software Foundation, Inc. + Contributed by Richard Henderson <rth@redhat.com>. + + This file is part of the GNU Atomic Library (libatomic). + + Libatomic is free software; you can redistribute it and/or modify it + under the terms of the GNU General Public License as published by + the Free Software Foundation; either version 3 of the License, or + (at your option) any later version. + + Libatomic is distributed in the hope that it will be useful, but WITHOUT ANY + WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS + FOR A PARTICULAR PURPOSE. See the GNU General Public License for + more details. + + Under Section 7 of GPL version 3, you are granted additional + permissions described in the GCC Runtime Library Exception, version + 3.1, as published by the Free Software Foundation. + + You should have received a copy of the GNU General Public License and + a copy of the GCC Runtime Library Exception along with this program; + see the files COPYING3 and COPYING.RUNTIME respectively. If not, see + <http://www.gnu.org/licenses/>. */ + +#include "libatomic_i.h" + + +/* If we natively support the exchange, and if we're unconcerned with extra + barriers (e.g. fully in-order cpu for which barriers are a nop), then + go ahead and expand the operation inline. */ +#if !defined(WANT_SPECIALCASE_RELAXED) && !defined(__OPTIMIZE_SIZE__) +# define EXACT_INLINE(N) \ + if (C2(HAVE_ATOMIC_EXCHANGE_,N)) \ + { \ + *PTR(N,rptr) = __atomic_exchange_n \ + (PTR(N,mptr), *PTR(N,vptr), smodel); \ + return; \ + } +#else +# define EXACT_INLINE(N) +#endif + + +#define EXACT(N) \ + do { \ + if (!C2(HAVE_INT,N)) break; \ + if ((uintptr_t)mptr & (N - 1)) break; \ + EXACT_INLINE (N); \ + *PTR(N,rptr) = C3(local_,exchange_,N) \ + (PTR(N,mptr), *PTR(N,vptr), smodel); \ + return; \ + } while (0) + + +#define LARGER(N) \ + do { \ + if (!C2(HAVE_INT,N)) break; \ + if (!C2(MAYBE_HAVE_ATOMIC_CAS_,N)) break; \ + r = (uintptr_t)mptr & (N - 1); \ + a = (uintptr_t)mptr & -N; \ + if (r + n <= N) \ + { \ + pre_barrier (smodel); \ + u.C2(i,N) = *PTR(N,a); \ + do { \ + v = u; \ + memcpy (v.b + r, vptr, n); \ + } while (!(C2(HAVE_ATOMIC_CAS_,N) \ + ? __atomic_compare_exchange_n (PTR(N,a), \ + &u.C2(i,N), v.C2(i,N), true, \ + __ATOMIC_RELAXED, __ATOMIC_RELAXED) \ + : C3(local_,compare_exchange_,N) (PTR(N,a), \ + &u.C2(i,N), v.C2(i,N), \ + __ATOMIC_RELAXED, __ATOMIC_RELAXED))); \ + goto Lfinish; \ + } \ + } while (0) + + +static void __attribute__((noinline)) +libat_exchange_large_inplace (size_t n, void *mptr, void *vptr) +{ +#define BUF 1024 + + char temp[BUF]; + size_t i = 0; + + for (i = 0; n >= BUF; i += BUF, n -= BUF) + { + memcpy (temp, mptr + i, BUF); + memcpy (mptr + i, vptr + i, BUF); + memcpy (vptr + i, temp, BUF); + } + if (n > 0) + { + memcpy (temp, mptr + i, n); + memcpy (mptr + i, vptr + i, n); + memcpy (vptr + i, temp, n); + } + +#undef BUF +} + +void +libat_exchange (size_t n, void *mptr, void *vptr, void *rptr, int smodel) +{ + union max_size_u u, v; + uintptr_t r, a; + + switch (n) + { + case 0: return; + case 1: EXACT(1); goto L4; + case 2: EXACT(2); goto L4; + case 4: EXACT(4); goto L8; + case 8: EXACT(8); goto L16; + case 16: EXACT(16); break; + + case 3: L4: LARGER(4); /* FALLTHRU */ + case 5 ... 7: L8: LARGER(8); /* FALLTHRU */ + case 9 ... 15: L16: LARGER(16); break; + + Lfinish: + post_barrier (smodel); + memcpy (rptr, u.b + r, n); + return; + } + + pre_seq_barrier (smodel); + libat_lock_n (mptr, n); + + if (vptr != rptr) + { + memcpy (rptr, mptr, n); + memcpy (mptr, vptr, n); + } + else + libat_exchange_large_inplace (n, mptr, vptr); + + libat_unlock_n (mptr, n); + post_seq_barrier (smodel); +} + +EXPORT_ALIAS (exchange); |