forked from Minki/linux
d4d9eab4ad
This may fix a reported bug where an R_TILEGX_64 in a module was not pointing to an aligned address. Reported-by: Simon Marchi <simon.marchi@polymtl.ca> Signed-off-by: Chris Metcalf <cmetcalf@tilera.com>
197 lines
4.9 KiB
C
197 lines
4.9 KiB
C
/*
|
|
* Copyright 2010 Tilera Corporation. All Rights Reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation, version 2.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
|
|
* NON INFRINGEMENT. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* These routines make two important assumptions:
|
|
*
|
|
* 1. atomic_t is really an int and can be freely cast back and forth
|
|
* (validated in __init_atomic_per_cpu).
|
|
*
|
|
* 2. userspace uses sys_cmpxchg() for all atomic operations, thus using
|
|
* the same locking convention that all the kernel atomic routines use.
|
|
*/
|
|
|
|
#ifndef _ASM_TILE_FUTEX_H
|
|
#define _ASM_TILE_FUTEX_H
|
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
#include <linux/futex.h>
|
|
#include <linux/uaccess.h>
|
|
#include <linux/errno.h>
|
|
#include <asm/atomic.h>
|
|
|
|
/*
|
|
* Support macros for futex operations. Do not use these macros directly.
|
|
* They assume "ret", "val", "oparg", and "uaddr" in the lexical context.
|
|
* __futex_cmpxchg() additionally assumes "oldval".
|
|
*/
|
|
|
|
#ifdef __tilegx__
|
|
|
|
#define __futex_asm(OP) \
|
|
asm("1: {" #OP " %1, %3, %4; movei %0, 0 }\n" \
|
|
".pushsection .fixup,\"ax\"\n" \
|
|
"0: { movei %0, %5; j 9f }\n" \
|
|
".section __ex_table,\"a\"\n" \
|
|
".align 8\n" \
|
|
".quad 1b, 0b\n" \
|
|
".popsection\n" \
|
|
"9:" \
|
|
: "=r" (ret), "=r" (val), "+m" (*(uaddr)) \
|
|
: "r" (uaddr), "r" (oparg), "i" (-EFAULT))
|
|
|
|
#define __futex_set() __futex_asm(exch4)
|
|
#define __futex_add() __futex_asm(fetchadd4)
|
|
#define __futex_or() __futex_asm(fetchor4)
|
|
#define __futex_andn() ({ oparg = ~oparg; __futex_asm(fetchand4); })
|
|
#define __futex_cmpxchg() \
|
|
({ __insn_mtspr(SPR_CMPEXCH_VALUE, oldval); __futex_asm(cmpexch4); })
|
|
|
|
#define __futex_xor() \
|
|
({ \
|
|
u32 oldval, n = oparg; \
|
|
if ((ret = __get_user(oldval, uaddr)) == 0) { \
|
|
do { \
|
|
oparg = oldval ^ n; \
|
|
__futex_cmpxchg(); \
|
|
} while (ret == 0 && oldval != val); \
|
|
} \
|
|
})
|
|
|
|
/* No need to prefetch, since the atomic ops go to the home cache anyway. */
|
|
#define __futex_prolog()
|
|
|
|
#else
|
|
|
|
#define __futex_call(FN) \
|
|
{ \
|
|
struct __get_user gu = FN((u32 __force *)uaddr, lock, oparg); \
|
|
val = gu.val; \
|
|
ret = gu.err; \
|
|
}
|
|
|
|
#define __futex_set() __futex_call(__atomic_xchg)
|
|
#define __futex_add() __futex_call(__atomic_xchg_add)
|
|
#define __futex_or() __futex_call(__atomic_or)
|
|
#define __futex_andn() __futex_call(__atomic_andn)
|
|
#define __futex_xor() __futex_call(__atomic_xor)
|
|
|
|
#define __futex_cmpxchg() \
|
|
{ \
|
|
struct __get_user gu = __atomic_cmpxchg((u32 __force *)uaddr, \
|
|
lock, oldval, oparg); \
|
|
val = gu.val; \
|
|
ret = gu.err; \
|
|
}
|
|
|
|
/*
|
|
* Find the lock pointer for the atomic calls to use, and issue a
|
|
* prefetch to the user address to bring it into cache. Similar to
|
|
* __atomic_setup(), but we can't do a read into the L1 since it might
|
|
* fault; instead we do a prefetch into the L2.
|
|
*/
|
|
#define __futex_prolog() \
|
|
int *lock; \
|
|
__insn_prefetch(uaddr); \
|
|
lock = __atomic_hashed_lock((int __force *)uaddr)
|
|
#endif
|
|
|
|
static inline int futex_atomic_op_inuser(int encoded_op, u32 __user *uaddr)
|
|
{
|
|
int op = (encoded_op >> 28) & 7;
|
|
int cmp = (encoded_op >> 24) & 15;
|
|
int oparg = (encoded_op << 8) >> 20;
|
|
int cmparg = (encoded_op << 20) >> 20;
|
|
int uninitialized_var(val), ret;
|
|
|
|
__futex_prolog();
|
|
|
|
/* The 32-bit futex code makes this assumption, so validate it here. */
|
|
BUILD_BUG_ON(sizeof(atomic_t) != sizeof(int));
|
|
|
|
if (encoded_op & (FUTEX_OP_OPARG_SHIFT << 28))
|
|
oparg = 1 << oparg;
|
|
|
|
if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32)))
|
|
return -EFAULT;
|
|
|
|
pagefault_disable();
|
|
switch (op) {
|
|
case FUTEX_OP_SET:
|
|
__futex_set();
|
|
break;
|
|
case FUTEX_OP_ADD:
|
|
__futex_add();
|
|
break;
|
|
case FUTEX_OP_OR:
|
|
__futex_or();
|
|
break;
|
|
case FUTEX_OP_ANDN:
|
|
__futex_andn();
|
|
break;
|
|
case FUTEX_OP_XOR:
|
|
__futex_xor();
|
|
break;
|
|
default:
|
|
ret = -ENOSYS;
|
|
break;
|
|
}
|
|
pagefault_enable();
|
|
|
|
if (!ret) {
|
|
switch (cmp) {
|
|
case FUTEX_OP_CMP_EQ:
|
|
ret = (val == cmparg);
|
|
break;
|
|
case FUTEX_OP_CMP_NE:
|
|
ret = (val != cmparg);
|
|
break;
|
|
case FUTEX_OP_CMP_LT:
|
|
ret = (val < cmparg);
|
|
break;
|
|
case FUTEX_OP_CMP_GE:
|
|
ret = (val >= cmparg);
|
|
break;
|
|
case FUTEX_OP_CMP_LE:
|
|
ret = (val <= cmparg);
|
|
break;
|
|
case FUTEX_OP_CMP_GT:
|
|
ret = (val > cmparg);
|
|
break;
|
|
default:
|
|
ret = -ENOSYS;
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static inline int futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr,
|
|
u32 oldval, u32 oparg)
|
|
{
|
|
int ret, val;
|
|
|
|
__futex_prolog();
|
|
|
|
if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32)))
|
|
return -EFAULT;
|
|
|
|
__futex_cmpxchg();
|
|
|
|
*uval = val;
|
|
return ret;
|
|
}
|
|
|
|
#endif /* !__ASSEMBLY__ */
|
|
|
|
#endif /* _ASM_TILE_FUTEX_H */
|