URL
https://opencores.org/ocsvn/openrisc/openrisc/trunk
Subversion Repositories openrisc
[/] [openrisc/] [trunk/] [gnu-dev/] [or1k-gcc/] [libjava/] [sysdep/] [powerpc/] [locks.h] - Rev 768
Go to most recent revision | Compare with Previous | Blame | View Log
// locks.h - Thread synchronization primitives. PowerPC implementation. /* Copyright (C) 2002,2008 Free Software Foundation This file is part of libgcj. This software is copyrighted work licensed under the terms of the Libgcj License. Please consult the file "LIBGCJ_LICENSE" for details. */ #ifndef __SYSDEP_LOCKS_H__ #define __SYSDEP_LOCKS_H__ #ifdef __LP64__ #define _LARX "ldarx " #define _STCX "stdcx. " #else #define _LARX "lwarx " #ifdef __PPC405__ #define _STCX "sync; stwcx. " #else #define _STCX "stwcx. " #endif #endif typedef size_t obj_addr_t; /* Integer type big enough for object */ /* address. */ inline static bool compare_and_swap (volatile obj_addr_t *addr, obj_addr_t old, obj_addr_t new_val) { obj_addr_t ret; __asm__ __volatile__ ( " " _LARX "%0,0,%1 \n" " xor. %0,%3,%0\n" " bne $+12\n" " " _STCX "%2,0,%1\n" " bne- $-16\n" : "=&r" (ret) : "r" (addr), "r" (new_val), "r" (old) : "cr0", "memory"); /* This version of __compare_and_swap is to be used when acquiring a lock, so we don't need to worry about whether other memory operations have completed, but we do need to be sure that any loads after this point really occur after we have acquired the lock. */ __asm__ __volatile__ ("isync" : : : "memory"); return ret == 0; } inline static void release_set (volatile obj_addr_t *addr, obj_addr_t new_val) { __asm__ __volatile__ ("sync" : : : "memory"); *addr = new_val; } inline static bool compare_and_swap_release (volatile obj_addr_t *addr, obj_addr_t old, obj_addr_t new_val) { obj_addr_t ret; __asm__ __volatile__ ("sync" : : : "memory"); __asm__ __volatile__ ( " " _LARX "%0,0,%1 \n" " xor. %0,%3,%0\n" " bne $+12\n" " " _STCX "%2,0,%1\n" " bne- $-16\n" : "=&r" (ret) : "r" (addr), "r" (new_val), "r" (old) : "cr0", "memory"); return ret == 0; } // Ensure that subsequent instructions do not execute on stale // data that was loaded from memory before the barrier. inline static void read_barrier () { __asm__ __volatile__ ("isync" : : : "memory"); } // Ensure that prior stores to memory are completed with respect to other // processors. inline static void write_barrier () { __asm__ __volatile__ ("sync" : : : "memory"); } #endif
Go to most recent revision | Compare with Previous | Blame | View Log