Skip to content

Commit

Permalink
asm-generic: unaligned: remove byteshift helpers
Browse files Browse the repository at this point in the history
In theory, compilers should be able to work this out themselves so we
can use a simpler version based on the swab() helpers.

I have verified that this works on all supported compiler versions
(gcc-4.9 and up, clang-10 and up). Looking at the object code produced by
gcc-11, I found that the impact is mostly a change in inlining decisions
that lead to slightly larger code.

In other cases, this version produces explicit byte swaps in place of
separate byte access, or comparing against pre-swapped constants.

While the source code is clearly simpler, I have not seen an indication
of the new version actually producing better code on Arm, so maybe
we want to skip this after all. From what I can tell, gcc recognizes
the byteswap pattern in the byteshift.h header and can turn it into
explicit instructions, but it does not turn a __builtin_bswap32() back
into individual bytes when that would result in better output, e.g.
when storing a byte-reversed constant.

Suggested-by: Linus Torvalds <torvalds@linux-foundation.org>
Signed-off-by: Arnd Bergmann <arnd@arndb.de>
  • Loading branch information
arndb committed May 10, 2021
1 parent f12d3ff commit 0652035
Show file tree
Hide file tree
Showing 6 changed files with 60 additions and 146 deletions.
2 changes: 0 additions & 2 deletions arch/arm/include/asm/unaligned.h
Original file line number Diff line number Diff line change
Expand Up @@ -10,13 +10,11 @@

#if defined(__LITTLE_ENDIAN)
# include <linux/unaligned/le_struct.h>
# include <linux/unaligned/be_byteshift.h>
# include <linux/unaligned/generic.h>
# define get_unaligned __get_unaligned_le
# define put_unaligned __put_unaligned_le
#elif defined(__BIG_ENDIAN)
# include <linux/unaligned/be_struct.h>
# include <linux/unaligned/le_byteshift.h>
# include <linux/unaligned/generic.h>
# define get_unaligned __get_unaligned_be
# define put_unaligned __put_unaligned_be
Expand Down
2 changes: 0 additions & 2 deletions include/asm-generic/unaligned.h
Original file line number Diff line number Diff line change
Expand Up @@ -16,15 +16,13 @@
#if defined(__LITTLE_ENDIAN)
# ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
# include <linux/unaligned/le_struct.h>
# include <linux/unaligned/be_byteshift.h>
# endif
# include <linux/unaligned/generic.h>
# define get_unaligned __get_unaligned_le
# define put_unaligned __put_unaligned_le
#elif defined(__BIG_ENDIAN)
# ifndef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
# include <linux/unaligned/be_struct.h>
# include <linux/unaligned/le_byteshift.h>
# endif
# include <linux/unaligned/generic.h>
# define get_unaligned __get_unaligned_be
Expand Down
71 changes: 0 additions & 71 deletions include/linux/unaligned/be_byteshift.h

This file was deleted.

30 changes: 30 additions & 0 deletions include/linux/unaligned/be_struct.h
Original file line number Diff line number Diff line change
Expand Up @@ -34,4 +34,34 @@ static inline void put_unaligned_be64(u64 val, void *p)
__put_unaligned_cpu64(val, p);
}

static inline u16 get_unaligned_le16(const void *p)
{
return swab16(__get_unaligned_cpu16((const u8 *)p));
}

static inline u32 get_unaligned_le32(const void *p)
{
return swab32(__get_unaligned_cpu32((const u8 *)p));
}

static inline u64 get_unaligned_le64(const void *p)
{
return swab64(__get_unaligned_cpu64((const u8 *)p));
}

static inline void put_unaligned_le16(u16 val, void *p)
{
__put_unaligned_cpu16(swab16(val), p);
}

static inline void put_unaligned_le32(u32 val, void *p)
{
__put_unaligned_cpu32(swab32(val), p);
}

static inline void put_unaligned_le64(u64 val, void *p)
{
__put_unaligned_cpu64(swab64(val), p);
}

#endif /* _LINUX_UNALIGNED_BE_STRUCT_H */
71 changes: 0 additions & 71 deletions include/linux/unaligned/le_byteshift.h

This file was deleted.

30 changes: 30 additions & 0 deletions include/linux/unaligned/le_struct.h
Original file line number Diff line number Diff line change
Expand Up @@ -34,4 +34,34 @@ static inline void put_unaligned_le64(u64 val, void *p)
__put_unaligned_cpu64(val, p);
}

static inline u16 get_unaligned_be16(const void *p)
{
return swab16(__get_unaligned_cpu16((const u8 *)p));
}

static inline u32 get_unaligned_be32(const void *p)
{
return swab32(__get_unaligned_cpu32((const u8 *)p));
}

static inline u64 get_unaligned_be64(const void *p)
{
return swab64(__get_unaligned_cpu64((const u8 *)p));
}

static inline void put_unaligned_be16(u16 val, void *p)
{
__put_unaligned_cpu16(swab16(val), p);
}

static inline void put_unaligned_be32(u32 val, void *p)
{
__put_unaligned_cpu32(swab32(val), p);
}

static inline void put_unaligned_be64(u64 val, void *p)
{
__put_unaligned_cpu64(swab64(val), p);
}

#endif /* _LINUX_UNALIGNED_LE_STRUCT_H */

0 comments on commit 0652035

Please sign in to comment.