|  | #ifndef _ASM_PARISC_FUTEX_H | 
|  | #define _ASM_PARISC_FUTEX_H | 
|  |  | 
|  | #ifdef __KERNEL__ | 
|  |  | 
|  | #include <linux/futex.h> | 
|  | #include <linux/uaccess.h> | 
|  | #include <asm/atomic.h> | 
|  | #include <asm/errno.h> | 
|  |  | 
|  | /* The following has to match the LWS code in syscall.S.  We have | 
|  | sixteen four-word locks. */ | 
|  |  | 
|  | static inline void | 
|  | _futex_spin_lock_irqsave(u32 __user *uaddr, unsigned long int *flags) | 
|  | { | 
|  | extern u32 lws_lock_start[]; | 
|  | long index = ((long)uaddr & 0xf0) >> 2; | 
|  | arch_spinlock_t *s = (arch_spinlock_t *)&lws_lock_start[index]; | 
|  | local_irq_save(*flags); | 
|  | arch_spin_lock(s); | 
|  | } | 
|  |  | 
|  | static inline void | 
|  | _futex_spin_unlock_irqrestore(u32 __user *uaddr, unsigned long int *flags) | 
|  | { | 
|  | extern u32 lws_lock_start[]; | 
|  | long index = ((long)uaddr & 0xf0) >> 2; | 
|  | arch_spinlock_t *s = (arch_spinlock_t *)&lws_lock_start[index]; | 
|  | arch_spin_unlock(s); | 
|  | local_irq_restore(*flags); | 
|  | } | 
|  |  | 
|  | static inline int | 
|  | futex_atomic_op_inuser (int encoded_op, u32 __user *uaddr) | 
|  | { | 
|  | unsigned long int flags; | 
|  | u32 val; | 
|  | int op = (encoded_op >> 28) & 7; | 
|  | int cmp = (encoded_op >> 24) & 15; | 
|  | int oparg = (encoded_op << 8) >> 20; | 
|  | int cmparg = (encoded_op << 20) >> 20; | 
|  | int oldval = 0, ret; | 
|  | if (encoded_op & (FUTEX_OP_OPARG_SHIFT << 28)) | 
|  | oparg = 1 << oparg; | 
|  |  | 
|  | if (!access_ok(VERIFY_WRITE, uaddr, sizeof(*uaddr))) | 
|  | return -EFAULT; | 
|  |  | 
|  | pagefault_disable(); | 
|  |  | 
|  | _futex_spin_lock_irqsave(uaddr, &flags); | 
|  |  | 
|  | switch (op) { | 
|  | case FUTEX_OP_SET: | 
|  | /* *(int *)UADDR2 = OPARG; */ | 
|  | ret = get_user(oldval, uaddr); | 
|  | if (!ret) | 
|  | ret = put_user(oparg, uaddr); | 
|  | break; | 
|  | case FUTEX_OP_ADD: | 
|  | /* *(int *)UADDR2 += OPARG; */ | 
|  | ret = get_user(oldval, uaddr); | 
|  | if (!ret) { | 
|  | val = oldval + oparg; | 
|  | ret = put_user(val, uaddr); | 
|  | } | 
|  | break; | 
|  | case FUTEX_OP_OR: | 
|  | /* *(int *)UADDR2 |= OPARG; */ | 
|  | ret = get_user(oldval, uaddr); | 
|  | if (!ret) { | 
|  | val = oldval | oparg; | 
|  | ret = put_user(val, uaddr); | 
|  | } | 
|  | break; | 
|  | case FUTEX_OP_ANDN: | 
|  | /* *(int *)UADDR2 &= ~OPARG; */ | 
|  | ret = get_user(oldval, uaddr); | 
|  | if (!ret) { | 
|  | val = oldval & ~oparg; | 
|  | ret = put_user(val, uaddr); | 
|  | } | 
|  | break; | 
|  | case FUTEX_OP_XOR: | 
|  | /* *(int *)UADDR2 ^= OPARG; */ | 
|  | ret = get_user(oldval, uaddr); | 
|  | if (!ret) { | 
|  | val = oldval ^ oparg; | 
|  | ret = put_user(val, uaddr); | 
|  | } | 
|  | break; | 
|  | default: | 
|  | ret = -ENOSYS; | 
|  | } | 
|  |  | 
|  | _futex_spin_unlock_irqrestore(uaddr, &flags); | 
|  |  | 
|  | pagefault_enable(); | 
|  |  | 
|  | if (!ret) { | 
|  | switch (cmp) { | 
|  | case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break; | 
|  | case FUTEX_OP_CMP_NE: ret = (oldval != cmparg); break; | 
|  | case FUTEX_OP_CMP_LT: ret = (oldval < cmparg); break; | 
|  | case FUTEX_OP_CMP_GE: ret = (oldval >= cmparg); break; | 
|  | case FUTEX_OP_CMP_LE: ret = (oldval <= cmparg); break; | 
|  | case FUTEX_OP_CMP_GT: ret = (oldval > cmparg); break; | 
|  | default: ret = -ENOSYS; | 
|  | } | 
|  | } | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Non-atomic version */ | 
|  | static inline int | 
|  | futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, | 
|  | u32 oldval, u32 newval) | 
|  | { | 
|  | int ret; | 
|  | u32 val; | 
|  | unsigned long flags; | 
|  |  | 
|  | /* futex.c wants to do a cmpxchg_inatomic on kernel NULL, which is | 
|  | * our gateway page, and causes no end of trouble... | 
|  | */ | 
|  | if (segment_eq(KERNEL_DS, get_fs()) && !uaddr) | 
|  | return -EFAULT; | 
|  |  | 
|  | if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32))) | 
|  | return -EFAULT; | 
|  |  | 
|  | /* HPPA has no cmpxchg in hardware and therefore the | 
|  | * best we can do here is use an array of locks. The | 
|  | * lock selected is based on a hash of the userspace | 
|  | * address. This should scale to a couple of CPUs. | 
|  | */ | 
|  |  | 
|  | _futex_spin_lock_irqsave(uaddr, &flags); | 
|  |  | 
|  | ret = get_user(val, uaddr); | 
|  |  | 
|  | if (!ret && val == oldval) | 
|  | ret = put_user(newval, uaddr); | 
|  |  | 
|  | *uval = val; | 
|  |  | 
|  | _futex_spin_unlock_irqrestore(uaddr, &flags); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | #endif /*__KERNEL__*/ | 
|  | #endif /*_ASM_PARISC_FUTEX_H*/ |