/* $OpenBSD: _atomic_lock.c,v 1.3 2006/01/05 22:33:24 marc Exp $ */ /* * Atomic lock for powerpc */ #include int _atomic_lock(volatile _spinlock_lock_t *lock) { _spinlock_lock_t old; __asm__("1: lwarx 0,0,%1 \n" " stwcx. %2,0,%1 \n" " bne- 1b \n" " mr %0, 0 \n" : "=r" (old), "=r" (lock) : "r" (_SPINLOCK_LOCKED), "1" (lock) : "0" ); return (old != _SPINLOCK_UNLOCKED); /* * Dale says: * Side note. to prevent two processes from accessing * the same address with the lwarx in one instrution * and the stwcx in another process, the current powerpc * kernel uses a stwcx instruction without the corresponding * lwarx which causes any reservation of a process * to be removed. if a context switch occurs * between the two accesses the store will not occur * and the condition code will cause it to loop. If on * a dual processor machine, the reserve will cause * appropriate bus cycle accesses to notify other * processors. */ }