|
|
@ -131,6 +131,15 @@ static inline void arch_write_lock(arch_rwlock_t *rw) |
|
|
|
*(volatile __u32 *)&lp->lock = ~0U; |
|
|
|
*(volatile __u32 *)&lp->lock = ~0U; |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static void inline arch_write_unlock(arch_rwlock_t *lock) |
|
|
|
|
|
|
|
{ |
|
|
|
|
|
|
|
__asm__ __volatile__( |
|
|
|
|
|
|
|
" st %%g0, [%0]" |
|
|
|
|
|
|
|
: /* no outputs */ |
|
|
|
|
|
|
|
: "r" (lock) |
|
|
|
|
|
|
|
: "memory"); |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
static inline int arch_write_trylock(arch_rwlock_t *rw) |
|
|
|
static inline int arch_write_trylock(arch_rwlock_t *rw) |
|
|
|
{ |
|
|
|
{ |
|
|
|
unsigned int val; |
|
|
|
unsigned int val; |
|
|
@ -175,8 +184,6 @@ static inline int __arch_read_trylock(arch_rwlock_t *rw) |
|
|
|
res; \
|
|
|
|
res; \
|
|
|
|
}) |
|
|
|
}) |
|
|
|
|
|
|
|
|
|
|
|
#define arch_write_unlock(rw) do { (rw)->lock = 0; } while(0) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#define arch_spin_lock_flags(lock, flags) arch_spin_lock(lock) |
|
|
|
#define arch_spin_lock_flags(lock, flags) arch_spin_lock(lock) |
|
|
|
#define arch_read_lock_flags(rw, flags) arch_read_lock(rw) |
|
|
|
#define arch_read_lock_flags(rw, flags) arch_read_lock(rw) |
|
|
|
#define arch_write_lock_flags(rw, flags) arch_write_lock(rw) |
|
|
|
#define arch_write_lock_flags(rw, flags) arch_write_lock(rw) |
|
|
|