diff options
author | Glauber Costa <gcosta@redhat.com> | 2008-06-25 14:53:41 -0300 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2008-07-09 09:14:28 +0200 |
commit | 8bc7de0c5dc0a5d3bcdc04bac6de0c799f91c5e4 (patch) | |
tree | c1f8dfbe35c408f41d89d345a0f8bf03d2af5dd9 /include/asm-x86/uaccess_32.h | |
parent | 8cb834e99f44bd56409b794504ae2b170675fc92 (diff) | |
download | op-kernel-dev-8bc7de0c5dc0a5d3bcdc04bac6de0c799f91c5e4.zip op-kernel-dev-8bc7de0c5dc0a5d3bcdc04bac6de0c799f91c5e4.tar.gz |
x86: put movsl_mask into uaccess.h.
x86_64 does not need it, but it won't have X86_INTEL_USERCOPY
defined either.
Signed-off-by: Glauber Costa <gcosta@redhat.com>
Signed-off-by: H. Peter Anvin <hpa@zytor.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'include/asm-x86/uaccess_32.h')
-rw-r--r-- | include/asm-x86/uaccess_32.h | 9 |
1 files changed, 0 insertions, 9 deletions
diff --git a/include/asm-x86/uaccess_32.h b/include/asm-x86/uaccess_32.h index d3b5bf8..3467749 100644 --- a/include/asm-x86/uaccess_32.h +++ b/include/asm-x86/uaccess_32.h @@ -11,15 +11,6 @@ #include <asm/asm.h> #include <asm/page.h> -/* - * movsl can be slow when source and dest are not both 8-byte aligned - */ -#ifdef CONFIG_X86_INTEL_USERCOPY -extern struct movsl_mask { - int mask; -} ____cacheline_aligned_in_smp movsl_mask; -#endif - unsigned long __must_check __copy_to_user_ll (void __user *to, const void *from, unsigned long n); unsigned long __must_check __copy_from_user_ll |