Lines Matching refs:n

25 static inline int __movsl_is_ok(unsigned long a1, unsigned long a2, unsigned long n)  in __movsl_is_ok()  argument
28 if (n >= 64 && ((a1 ^ a2) & movsl_mask.mask)) in __movsl_is_ok()
33 #define movsl_is_ok(a1, a2, n) \ argument
34 __movsl_is_ok((unsigned long)(a1), (unsigned long)(a2), (n))
71 clear_user(void __user *to, unsigned long n) in clear_user() argument
74 if (access_ok(VERIFY_WRITE, to, n)) in clear_user()
75 __do_clear_user(to, n); in clear_user()
76 return n; in clear_user()
92 __clear_user(void __user *to, unsigned long n) in __clear_user() argument
94 __do_clear_user(to, n); in __clear_user()
95 return n; in __clear_user()
571 unsigned long n) in __copy_to_user_ll() argument
574 if (movsl_is_ok(to, from, n)) in __copy_to_user_ll()
575 __copy_user(to, from, n); in __copy_to_user_ll()
577 n = __copy_user_intel(to, from, n); in __copy_to_user_ll()
579 return n; in __copy_to_user_ll()
584 unsigned long n) in __copy_from_user_ll() argument
587 if (movsl_is_ok(to, from, n)) in __copy_from_user_ll()
588 __copy_user_zeroing(to, from, n); in __copy_from_user_ll()
590 n = __copy_user_zeroing_intel(to, from, n); in __copy_from_user_ll()
592 return n; in __copy_from_user_ll()
597 unsigned long n) in __copy_from_user_ll_nozero() argument
600 if (movsl_is_ok(to, from, n)) in __copy_from_user_ll_nozero()
601 __copy_user(to, from, n); in __copy_from_user_ll_nozero()
603 n = __copy_user_intel((void __user *)to, in __copy_from_user_ll_nozero()
604 (const void *)from, n); in __copy_from_user_ll_nozero()
606 return n; in __copy_from_user_ll_nozero()
611 unsigned long n) in __copy_from_user_ll_nocache() argument
615 if (n > 64 && cpu_has_xmm2) in __copy_from_user_ll_nocache()
616 n = __copy_user_zeroing_intel_nocache(to, from, n); in __copy_from_user_ll_nocache()
618 __copy_user_zeroing(to, from, n); in __copy_from_user_ll_nocache()
620 __copy_user_zeroing(to, from, n); in __copy_from_user_ll_nocache()
623 return n; in __copy_from_user_ll_nocache()
628 unsigned long n) in __copy_from_user_ll_nocache_nozero() argument
632 if (n > 64 && cpu_has_xmm2) in __copy_from_user_ll_nocache_nozero()
633 n = __copy_user_intel_nocache(to, from, n); in __copy_from_user_ll_nocache_nozero()
635 __copy_user(to, from, n); in __copy_from_user_ll_nocache_nozero()
637 __copy_user(to, from, n); in __copy_from_user_ll_nocache_nozero()
640 return n; in __copy_from_user_ll_nocache_nozero()
657 unsigned long _copy_to_user(void __user *to, const void *from, unsigned n) in _copy_to_user() argument
659 if (access_ok(VERIFY_WRITE, to, n)) in _copy_to_user()
660 n = __copy_to_user(to, from, n); in _copy_to_user()
661 return n; in _copy_to_user()
681 unsigned long _copy_from_user(void *to, const void __user *from, unsigned n) in _copy_from_user() argument
683 if (access_ok(VERIFY_READ, from, n)) in _copy_from_user()
684 n = __copy_from_user(to, from, n); in _copy_from_user()
686 memset(to, 0, n); in _copy_from_user()
687 return n; in _copy_from_user()