diff options
Diffstat (limited to 'target/linux')
-rw-r--r-- | target/linux/generic-2.6/patches-2.6.28/300-fix_byteorder_header.patch | 348 |
1 files changed, 348 insertions, 0 deletions
diff --git a/target/linux/generic-2.6/patches-2.6.28/300-fix_byteorder_header.patch b/target/linux/generic-2.6/patches-2.6.28/300-fix_byteorder_header.patch new file mode 100644 index 000000000..8a4e08853 --- /dev/null +++ b/target/linux/generic-2.6/patches-2.6.28/300-fix_byteorder_header.patch @@ -0,0 +1,348 @@ +glibc headers define both __LITTLE_ENDIAN and __BIG_ENDIAN +which was tripping the check in linux/byteorder.h. Let's +just stay out of userspace's way and use __KERN_{endian} + +The old implementation got away with it as it kept the two +implementations in different headers and conditionally included +the right one. The combined header does checks within each +function body and depends on only one being defined. + +Converted the arches in mainline that have already moved to the +new header, as the other arches merge the will need simlar +fixups. + +Signed-off-by: Harvey Harrison <harvey.harrison@gmail.com> +--- + arch/avr32/include/asm/byteorder.h | 2 +- + arch/mips/include/asm/byteorder.h | 4 +- + arch/sh/include/asm/byteorder.h | 4 +- + arch/sparc/include/asm/byteorder.h | 2 +- + include/linux/byteorder.h | 84 ++++++++++++++++++------------------ + 5 files changed, 48 insertions(+), 48 deletions(-) + +diff --git a/arch/avr32/include/asm/byteorder.h b/arch/avr32/include/asm/byteorder.h +index 8e3af02..b7d6dd1 100644 +--- a/arch/avr32/include/asm/byteorder.h ++++ b/arch/avr32/include/asm/byteorder.h +@@ -7,7 +7,7 @@ + #include <asm/types.h> + #include <linux/compiler.h> + +-#define __BIG_ENDIAN ++#define __KERN_BIG_ENDIAN + #define __SWAB_64_THRU_32__ + + #ifdef __CHECKER__ +diff --git a/arch/mips/include/asm/byteorder.h b/arch/mips/include/asm/byteorder.h +index 2988d29..8ad8a5b 100644 +--- a/arch/mips/include/asm/byteorder.h ++++ b/arch/mips/include/asm/byteorder.h +@@ -12,9 +12,9 @@ + #include <asm/types.h> + + #if defined(__MIPSEB__) +-# define __BIG_ENDIAN ++# define __KERN_BIG_ENDIAN + #elif defined(__MIPSEL__) +-# define __LITTLE_ENDIAN ++# define __KERN_LITTLE_ENDIAN + #else + # error "MIPS, but neither __MIPSEB__, nor __MIPSEL__???" + #endif +diff --git a/arch/sh/include/asm/byteorder.h b/arch/sh/include/asm/byteorder.h +index f5fa065..4aa5f1d 100644 +--- a/arch/sh/include/asm/byteorder.h ++++ b/arch/sh/include/asm/byteorder.h +@@ -9,9 +9,9 @@ + #include <linux/types.h> + + #ifdef __LITTLE_ENDIAN__ +-# define __LITTLE_ENDIAN ++# define __KERN_LITTLE_ENDIAN + #else +-# define __BIG_ENDIAN ++# define __KERN_BIG_ENDIAN + #endif + + #define __SWAB_64_THRU_32__ +diff --git a/arch/sparc/include/asm/byteorder.h b/arch/sparc/include/asm/byteorder.h +index 5a70f13..5b8347e 100644 +--- a/arch/sparc/include/asm/byteorder.h ++++ b/arch/sparc/include/asm/byteorder.h +@@ -4,7 +4,7 @@ + #include <asm/types.h> + #include <asm/asi.h> + +-#define __BIG_ENDIAN ++#define __KERN_BIG_ENDIAN + + #ifdef CONFIG_SPARC32 + #define __SWAB_64_THRU_32__ +diff --git a/include/linux/byteorder.h b/include/linux/byteorder.h +index 29f002d..3599fbc 100644 +--- a/include/linux/byteorder.h ++++ b/include/linux/byteorder.h +@@ -4,33 +4,33 @@ + #include <linux/types.h> + #include <linux/swab.h> + +-#if defined(__LITTLE_ENDIAN) && defined(__BIG_ENDIAN) ++#if defined(__KERN_LITTLE_ENDIAN) && defined(__KERN_BIG_ENDIAN) + # error Fix asm/byteorder.h to define one endianness + #endif + +-#if !defined(__LITTLE_ENDIAN) && !defined(__BIG_ENDIAN) ++#if !defined(__KERN_LITTLE_ENDIAN) && !defined(__KERN_BIG_ENDIAN) + # error Fix asm/byteorder.h to define arch endianness + #endif + +-#ifdef __LITTLE_ENDIAN +-# undef __LITTLE_ENDIAN +-# define __LITTLE_ENDIAN 1234 ++#ifdef __KERN_LITTLE_ENDIAN ++# ifndef __LITTLE_ENDIAN ++# define __LITTLE_ENDIAN 1234 ++# endif ++# ifndef __LITTLE_ENDIAN_BITFIELD ++# define __LITTLE_ENDIAN_BITFIELD ++# endif + #endif + +-#ifdef __BIG_ENDIAN +-# undef __BIG_ENDIAN +-# define __BIG_ENDIAN 4321 ++#ifdef __KERN_BIG_ENDIAN ++# ifndef __BIG_ENDIAN ++# define __BIG_ENDIAN 4321 ++# endif ++# ifndef __BIG_ENDIAN_BITFIELD ++# define __BIG_ENDIAN_BITFIELD ++# endif + #endif + +-#if defined(__LITTLE_ENDIAN) && !defined(__LITTLE_ENDIAN_BITFIELD) +-# define __LITTLE_ENDIAN_BITFIELD +-#endif +- +-#if defined(__BIG_ENDIAN) && !defined(__BIG_ENDIAN_BITFIELD) +-# define __BIG_ENDIAN_BITFIELD +-#endif +- +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + # define __le16_to_cpu(x) ((__force __u16)(__le16)(x)) + # define __le32_to_cpu(x) ((__force __u32)(__le32)(x)) + # define __le64_to_cpu(x) ((__force __u64)(__le64)(x)) +@@ -46,7 +46,7 @@ + # define __cpu_to_be64(x) ((__force __be64)__swab64(x)) + #endif + +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + # define __be16_to_cpu(x) ((__force __u16)(__be16)(x)) + # define __be32_to_cpu(x) ((__force __u32)(__be32)(x)) + # define __be64_to_cpu(x) ((__force __u64)(__be64)(x)) +@@ -87,91 +87,91 @@ + + static inline void __le16_to_cpus(__u16 *p) + { +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + __swab16s(p); + #endif + } + + static inline void __cpu_to_le16s(__u16 *p) + { +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + __swab16s(p); + #endif + } + + static inline void __le32_to_cpus(__u32 *p) + { +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + __swab32s(p); + #endif + } + + static inline void __cpu_to_le32s(__u32 *p) + { +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + __swab32s(p); + #endif + } + + static inline void __le64_to_cpus(__u64 *p) + { +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + __swab64s(p); + #endif + } + + static inline void __cpu_to_le64s(__u64 *p) + { +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + __swab64s(p); + #endif + } + + static inline void __be16_to_cpus(__u16 *p) + { +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + __swab16s(p); + #endif + } + + static inline void __cpu_to_be16s(__u16 *p) + { +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + __swab16s(p); + #endif + } + + static inline void __be32_to_cpus(__u32 *p) + { +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + __swab32s(p); + #endif + } + + static inline void __cpu_to_be32s(__u32 *p) + { +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + __swab32s(p); + #endif + } + + static inline void __be64_to_cpus(__u64 *p) + { +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + __swab64s(p); + #endif + } + + static inline void __cpu_to_be64s(__u64 *p) + { +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + __swab64s(p); + #endif + } + + static inline __u16 __le16_to_cpup(const __le16 *p) + { +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + return (__force __u16)*p; + #else + return __swab16p((__force __u16 *)p); +@@ -180,7 +180,7 @@ static inline __u16 __le16_to_cpup(const __le16 *p) + + static inline __u32 __le32_to_cpup(const __le32 *p) + { +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + return (__force __u32)*p; + #else + return __swab32p((__force __u32 *)p); +@@ -189,7 +189,7 @@ static inline __u32 __le32_to_cpup(const __le32 *p) + + static inline __u64 __le64_to_cpup(const __le64 *p) + { +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + return (__force __u64)*p; + #else + return __swab64p((__force __u64 *)p); +@@ -198,7 +198,7 @@ static inline __u64 __le64_to_cpup(const __le64 *p) + + static inline __le16 __cpu_to_le16p(const __u16 *p) + { +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + return (__force __le16)*p; + #else + return (__force __le16)__swab16p(p); +@@ -207,7 +207,7 @@ static inline __le16 __cpu_to_le16p(const __u16 *p) + + static inline __le32 __cpu_to_le32p(const __u32 *p) + { +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + return (__force __le32)*p; + #else + return (__force __le32)__swab32p(p); +@@ -216,7 +216,7 @@ static inline __le32 __cpu_to_le32p(const __u32 *p) + + static inline __le64 __cpu_to_le64p(const __u64 *p) + { +-#ifdef __LITTLE_ENDIAN ++#ifdef __KERN_LITTLE_ENDIAN + return (__force __le64)*p; + #else + return (__force __le64)__swab64p(p); +@@ -225,7 +225,7 @@ static inline __le64 __cpu_to_le64p(const __u64 *p) + + static inline __u16 __be16_to_cpup(const __be16 *p) + { +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + return (__force __u16)*p; + #else + return __swab16p((__force __u16 *)p); +@@ -234,7 +234,7 @@ static inline __u16 __be16_to_cpup(const __be16 *p) + + static inline __u32 __be32_to_cpup(const __be32 *p) + { +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + return (__force __u32)*p; + #else + return __swab32p((__force __u32 *)p); +@@ -243,7 +243,7 @@ static inline __u32 __be32_to_cpup(const __be32 *p) + + static inline __u64 __be64_to_cpup(const __be64 *p) + { +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + return (__force __u64)*p; + #else + return __swab64p((__force __u64 *)p); +@@ -252,7 +252,7 @@ static inline __u64 __be64_to_cpup(const __be64 *p) + + static inline __be16 __cpu_to_be16p(const __u16 *p) + { +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + return (__force __be16)*p; + #else + return (__force __be16)__swab16p(p); +@@ -261,7 +261,7 @@ static inline __be16 __cpu_to_be16p(const __u16 *p) + + static inline __be32 __cpu_to_be32p(const __u32 *p) + { +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + return (__force __be32)*p; + #else + return (__force __be32)__swab32p(p); +@@ -270,7 +270,7 @@ static inline __be32 __cpu_to_be32p(const __u32 *p) + + static inline __be64 __cpu_to_be64p(const __u64 *p) + { +-#ifdef __BIG_ENDIAN ++#ifdef __KERN_BIG_ENDIAN + return (__force __be64)*p; + #else + return (__force __be64)__swab64p(p); |