1 #ifndef _I386_BYTEORDER_H
2 #define _I386_BYTEORDER_H
5 #include <linux/compiler.h>
9 /* For avoiding bswap on i386 */
11 #include <linux/config.h>
14 static __inline__ __attribute_const__ __u32 ___arch__swab32(__u32 x)
16 #ifdef CONFIG_X86_BSWAP
17 __asm__("bswap %0" : "=r" (x) : "0" (x));
19 __asm__("xchgb %b0,%h0\n\t" /* swap lower bytes */
20 "rorl $16,%0\n\t" /* swap words */
21 "xchgb %b0,%h0" /* swap higher bytes */
28 /* gcc should generate this for open coded C now too. May be worth switching to
29 it because inline assembly cannot be scheduled. -AK */
30 static __inline__ __attribute_const__ __u16 ___arch__swab16(__u16 x)
32 __asm__("xchgb %b0,%h0" /* swap bytes */
39 static inline __u64 ___arch__swab64(__u64 val)
42 struct { __u32 a,b; } s;
46 #ifdef CONFIG_X86_BSWAP
47 asm("bswapl %0 ; bswapl %1 ; xchgl %0,%1"
48 : "=r" (v.s.a), "=r" (v.s.b)
49 : "0" (v.s.a), "1" (v.s.b));
51 v.s.a = ___arch__swab32(v.s.a);
52 v.s.b = ___arch__swab32(v.s.b);
53 asm("xchgl %0,%1" : "=r" (v.s.a), "=r" (v.s.b) : "0" (v.s.a), "1" (v.s.b));
58 #define __arch__swab64(x) ___arch__swab64(x)
59 #define __arch__swab32(x) ___arch__swab32(x)
60 #define __arch__swab16(x) ___arch__swab16(x)
62 #define __BYTEORDER_HAS_U64__
66 #include <linux/byteorder/little_endian.h>
68 #endif /* _I386_BYTEORDER_H */