URL
https://opencores.org/ocsvn/openrisc/openrisc/trunk
Subversion Repositories openrisc
[/] [openrisc/] [trunk/] [gnu-dev/] [or1k-gcc/] [libstdc++-v3/] [include/] [parallel/] [compatibility.h] - Rev 775
Go to most recent revision | Compare with Previous | Blame | View Log
// -*- C++ -*- // Copyright (C) 2007, 2008, 2009, 2010, 2012 Free Software Foundation, Inc. // // This file is part of the GNU ISO C++ Library. This library is free // software; you can redistribute it and/or modify it under the terms // of the GNU General Public License as published by the Free Software // Foundation; either version 3, or (at your option) any later // version. // This library is distributed in the hope that it will be useful, but // WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU // General Public License for more details. // Under Section 7 of GPL version 3, you are granted additional // permissions described in the GCC Runtime Library Exception, version // 3.1, as published by the Free Software Foundation. // You should have received a copy of the GNU General Public License and // a copy of the GCC Runtime Library Exception along with this program; // see the files COPYING3 and COPYING.RUNTIME respectively. If not, see // <http://www.gnu.org/licenses/>. /** @file parallel/compatibility.h * @brief Compatibility layer, mostly concerned with atomic operations. * This file is a GNU parallel extension to the Standard C++ Library. */ // Written by Felix Putze. #ifndef _GLIBCXX_PARALLEL_COMPATIBILITY_H #define _GLIBCXX_PARALLEL_COMPATIBILITY_H 1 #include <parallel/types.h> #include <parallel/base.h> #if defined(__SUNPRO_CC) && defined(__sparc) #include <sys/atomic.h> #endif #if !defined(_WIN32) || defined (__CYGWIN__) #include <sched.h> #endif #if defined(_MSC_VER) #include <Windows.h> #include <intrin.h> #undef max #undef min #endif #ifdef __MINGW32__ // Including <windows.h> will drag in all the windows32 names. Since // that can cause user code portability problems, we just declare the // one needed function here. extern "C" __attribute((dllimport)) void __attribute__((stdcall)) Sleep (unsigned long); #endif namespace __gnu_parallel { #if defined(__ICC) template<typename _MustBeInt = int> int32_t __faa32(int32_t* __x, int32_t __inc) { asm volatile("lock xadd %0,%1" : "=__r" (__inc), "=__m" (*__x) : "0" (__inc) : "memory"); return __inc; } #if defined(__x86_64) template<typename _MustBeInt = int> int64_t __faa64(int64_t* __x, int64_t __inc) { asm volatile("lock xadd %0,%1" : "=__r" (__inc), "=__m" (*__x) : "0" (__inc) : "memory"); return __inc; } #endif #endif // atomic functions only work on integers /** @brief Add a value to a variable, atomically. * * Implementation is heavily platform-dependent. * @param __ptr Pointer to a 32-bit signed integer. * @param __addend Value to add. */ inline int32_t __fetch_and_add_32(volatile int32_t* __ptr, int32_t __addend) { #if defined(__ICC) //x86 version return _InterlockedExchangeAdd((void*)__ptr, __addend); #elif defined(__ECC) //IA-64 version return _InterlockedExchangeAdd((void*)__ptr, __addend); #elif defined(__ICL) || defined(_MSC_VER) return _InterlockedExchangeAdd(reinterpret_cast<volatile long*>(__ptr), __addend); #elif defined(__GNUC__) return __atomic_fetch_add(__ptr, __addend, __ATOMIC_ACQ_REL); #elif defined(__SUNPRO_CC) && defined(__sparc) volatile int32_t __before, __after; do { __before = *__ptr; __after = __before + __addend; } while (atomic_cas_32((volatile unsigned int*)__ptr, __before, __after) != __before); return __before; #else //fallback, slow #pragma message("slow __fetch_and_add_32") int32_t __res; #pragma omp critical { __res = *__ptr; *(__ptr) += __addend; } return __res; #endif } /** @brief Add a value to a variable, atomically. * * Implementation is heavily platform-dependent. * @param __ptr Pointer to a 64-bit signed integer. * @param __addend Value to add. */ inline int64_t __fetch_and_add_64(volatile int64_t* __ptr, int64_t __addend) { #if defined(__ICC) && defined(__x86_64) //x86 version return __faa64<int>((int64_t*)__ptr, __addend); #elif defined(__ECC) //IA-64 version return _InterlockedExchangeAdd64((void*)__ptr, __addend); #elif defined(__ICL) || defined(_MSC_VER) #ifndef _WIN64 _GLIBCXX_PARALLEL_ASSERT(false); //not available in this case return 0; #else return _InterlockedExchangeAdd64(__ptr, __addend); #endif #elif defined(__GNUC__) && defined(__x86_64) return __atomic_fetch_add(__ptr, __addend, __ATOMIC_ACQ_REL); #elif defined(__GNUC__) && defined(__i386) && \ (defined(__i686) || defined(__pentium4) || defined(__athlon) \ || defined(__k8) || defined(__core2)) return __atomic_fetch_add(__ptr, __addend, __ATOMIC_ACQ_REL); #elif defined(__SUNPRO_CC) && defined(__sparc) volatile int64_t __before, __after; do { __before = *__ptr; __after = __before + __addend; } while (atomic_cas_64((volatile unsigned long long*)__ptr, __before, __after) != __before); return __before; #else //fallback, slow #if defined(__GNUC__) && defined(__i386) // XXX doesn'__t work with -march=native //#warning "please compile with -march=i686 or better" #endif #pragma message("slow __fetch_and_add_64") int64_t __res; #pragma omp critical { __res = *__ptr; *(__ptr) += __addend; } return __res; #endif } /** @brief Add a value to a variable, atomically. * * Implementation is heavily platform-dependent. * @param __ptr Pointer to a signed integer. * @param __addend Value to add. */ template<typename _Tp> inline _Tp __fetch_and_add(volatile _Tp* __ptr, _Tp __addend) { if (sizeof(_Tp) == sizeof(int32_t)) return (_Tp)__fetch_and_add_32((volatile int32_t*) __ptr, (int32_t)__addend); else if (sizeof(_Tp) == sizeof(int64_t)) return (_Tp)__fetch_and_add_64((volatile int64_t*) __ptr, (int64_t)__addend); else _GLIBCXX_PARALLEL_ASSERT(false); } #if defined(__ICC) template<typename _MustBeInt = int> inline int32_t __cas32(volatile int32_t* __ptr, int32_t __old, int32_t __nw) { int32_t __before; __asm__ __volatile__("lock; cmpxchgl %1,%2" : "=a"(__before) : "q"(__nw), "__m"(*(volatile long long*)(__ptr)), "0"(__old) : "memory"); return __before; } #if defined(__x86_64) template<typename _MustBeInt = int> inline int64_t __cas64(volatile int64_t *__ptr, int64_t __old, int64_t __nw) { int64_t __before; __asm__ __volatile__("lock; cmpxchgq %1,%2" : "=a"(__before) : "q"(__nw), "__m"(*(volatile long long*)(__ptr)), "0"(__old) : "memory"); return __before; } #endif #endif /** @brief Compare @c *__ptr and @c __comparand. If equal, let @c * *__ptr=__replacement and return @c true, return @c false otherwise. * * Implementation is heavily platform-dependent. * @param __ptr Pointer to 32-bit signed integer. * @param __comparand Compare value. * @param __replacement Replacement value. */ inline bool __compare_and_swap_32(volatile int32_t* __ptr, int32_t __comparand, int32_t __replacement) { #if defined(__ICC) //x86 version return _InterlockedCompareExchange((void*)__ptr, __replacement, __comparand) == __comparand; #elif defined(__ECC) //IA-64 version return _InterlockedCompareExchange((void*)__ptr, __replacement, __comparand) == __comparand; #elif defined(__ICL) || defined(_MSC_VER) return _InterlockedCompareExchange( reinterpret_cast<volatile long*>(__ptr), __replacement, __comparand) == __comparand; #elif defined(__GNUC__) return __atomic_compare_exchange_n(__ptr, &__comparand, __replacement, false, __ATOMIC_ACQ_REL, __ATOMIC_RELAXED); #elif defined(__SUNPRO_CC) && defined(__sparc) return atomic_cas_32((volatile unsigned int*)__ptr, __comparand, __replacement) == __comparand; #else #pragma message("slow __compare_and_swap_32") bool __res = false; #pragma omp critical { if (*__ptr == __comparand) { *__ptr = __replacement; __res = true; } } return __res; #endif } /** @brief Compare @c *__ptr and @c __comparand. If equal, let @c * *__ptr=__replacement and return @c true, return @c false otherwise. * * Implementation is heavily platform-dependent. * @param __ptr Pointer to 64-bit signed integer. * @param __comparand Compare value. * @param __replacement Replacement value. */ inline bool __compare_and_swap_64(volatile int64_t* __ptr, int64_t __comparand, int64_t __replacement) { #if defined(__ICC) && defined(__x86_64) //x86 version return __cas64<int>(__ptr, __comparand, __replacement) == __comparand; #elif defined(__ECC) //IA-64 version return _InterlockedCompareExchange64((void*)__ptr, __replacement, __comparand) == __comparand; #elif defined(__ICL) || defined(_MSC_VER) #ifndef _WIN64 _GLIBCXX_PARALLEL_ASSERT(false); //not available in this case return 0; #else return _InterlockedCompareExchange64(__ptr, __replacement, __comparand) == __comparand; #endif #elif defined(__GNUC__) && defined(__x86_64) return __atomic_compare_exchange_n(__ptr, &__comparand, __replacement, false, __ATOMIC_ACQ_REL, __ATOMIC_RELAXED); #elif defined(__GNUC__) && defined(__i386) && \ (defined(__i686) || defined(__pentium4) || defined(__athlon) \ || defined(__k8) || defined(__core2)) return __atomic_compare_exchange_n(__ptr, &__comparand, __replacement, false, __ATOMIC_ACQ_REL, __ATOMIC_RELAXED); #elif defined(__SUNPRO_CC) && defined(__sparc) return atomic_cas_64((volatile unsigned long long*)__ptr, __comparand, __replacement) == __comparand; #else #if defined(__GNUC__) && defined(__i386) // XXX -march=native //#warning "please compile with -march=i686 or better" #endif #pragma message("slow __compare_and_swap_64") bool __res = false; #pragma omp critical { if (*__ptr == __comparand) { *__ptr = __replacement; __res = true; } } return __res; #endif } /** @brief Compare @c *__ptr and @c __comparand. If equal, let @c * *__ptr=__replacement and return @c true, return @c false otherwise. * * Implementation is heavily platform-dependent. * @param __ptr Pointer to signed integer. * @param __comparand Compare value. * @param __replacement Replacement value. */ template<typename _Tp> inline bool __compare_and_swap(volatile _Tp* __ptr, _Tp __comparand, _Tp __replacement) { if (sizeof(_Tp) == sizeof(int32_t)) return __compare_and_swap_32((volatile int32_t*) __ptr, (int32_t)__comparand, (int32_t)__replacement); else if (sizeof(_Tp) == sizeof(int64_t)) return __compare_and_swap_64((volatile int64_t*) __ptr, (int64_t)__comparand, (int64_t)__replacement); else _GLIBCXX_PARALLEL_ASSERT(false); } /** @brief Yield the control to another thread, without waiting for the end to the time slice. */ inline void __yield() { #if defined (_WIN32) && !defined (__CYGWIN__) Sleep(0); #else sched_yield(); #endif } } // end namespace #endif /* _GLIBCXX_PARALLEL_COMPATIBILITY_H */
Go to most recent revision | Compare with Previous | Blame | View Log