13da9e200f
This reverts commitb3b77c8cae
, which was also totally broken (see commit0d2daf5cc8
that reverted the crc32 version of it). As reported by Stephen Rothwell, it causes problems on big-endian machines: > In file included from fs/jfs/jfs_types.h:33, > from fs/jfs/jfs_incore.h:26, > from fs/jfs/file.c:22: > fs/jfs/endian24.h:36:101: warning: "__LITTLE_ENDIAN" is not defined The kernel has never had that crazy "__BYTE_ORDER == __LITTLE_ENDIAN" model. It's not how we do things, and it isn't how we _should_ do things. So don't go there. Requested-by: Stephen Rothwell <sfr@canb.auug.org.au> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
77 lines
1.8 KiB
C
77 lines
1.8 KiB
C
#include <linux/kernel.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/types.h>
|
|
#include <asm/byteorder.h>
|
|
|
|
#define add_ssaaaa(sh, sl, ah, al, bh, bl) ({ \
|
|
unsigned int __sh = (ah); \
|
|
unsigned int __sl = (al); \
|
|
asm volatile( \
|
|
" alr %1,%3\n" \
|
|
" brc 12,0f\n" \
|
|
" ahi %0,1\n" \
|
|
"0: alr %0,%2" \
|
|
: "+&d" (__sh), "+d" (__sl) \
|
|
: "d" (bh), "d" (bl) : "cc"); \
|
|
(sh) = __sh; \
|
|
(sl) = __sl; \
|
|
})
|
|
|
|
#define sub_ddmmss(sh, sl, ah, al, bh, bl) ({ \
|
|
unsigned int __sh = (ah); \
|
|
unsigned int __sl = (al); \
|
|
asm volatile( \
|
|
" slr %1,%3\n" \
|
|
" brc 3,0f\n" \
|
|
" ahi %0,-1\n" \
|
|
"0: slr %0,%2" \
|
|
: "+&d" (__sh), "+d" (__sl) \
|
|
: "d" (bh), "d" (bl) : "cc"); \
|
|
(sh) = __sh; \
|
|
(sl) = __sl; \
|
|
})
|
|
|
|
/* a umul b = a mul b + (a>=2<<31) ? b<<32:0 + (b>=2<<31) ? a<<32:0 */
|
|
#define umul_ppmm(wh, wl, u, v) ({ \
|
|
unsigned int __wh = u; \
|
|
unsigned int __wl = v; \
|
|
asm volatile( \
|
|
" ltr 1,%0\n" \
|
|
" mr 0,%1\n" \
|
|
" jnm 0f\n" \
|
|
" alr 0,%1\n" \
|
|
"0: ltr %1,%1\n" \
|
|
" jnm 1f\n" \
|
|
" alr 0,%0\n" \
|
|
"1: lr %0,0\n" \
|
|
" lr %1,1\n" \
|
|
: "+d" (__wh), "+d" (__wl) \
|
|
: : "0", "1", "cc"); \
|
|
wh = __wh; \
|
|
wl = __wl; \
|
|
})
|
|
|
|
#ifdef __s390x__
|
|
#define udiv_qrnnd(q, r, n1, n0, d) \
|
|
do { unsigned long __n; \
|
|
unsigned int __r, __d; \
|
|
__n = ((unsigned long)(n1) << 32) + n0; \
|
|
__d = (d); \
|
|
(q) = __n / __d; \
|
|
(r) = __n % __d; \
|
|
} while (0)
|
|
#else
|
|
#define udiv_qrnnd(q, r, n1, n0, d) \
|
|
do { unsigned int __r; \
|
|
(q) = __udiv_qrnnd (&__r, (n1), (n0), (d)); \
|
|
(r) = __r; \
|
|
} while (0)
|
|
extern unsigned long __udiv_qrnnd (unsigned int *, unsigned int,
|
|
unsigned int , unsigned int);
|
|
#endif
|
|
|
|
#define UDIV_NEEDS_NORMALIZATION 0
|
|
|
|
#define abort() return 0
|
|
|
|
#define __BYTE_ORDER __BIG_ENDIAN
|