blob: 186d5618ead869eab2b5f833e9df8b234f796752 [file] [log] [blame]
/* ----------------------------------------------------------------------- *
*
* Copyright 1996-2017 The NASM Authors - All Rights Reserved
* See the file AUTHORS included with the NASM distribution for
* the specific copyright holders.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following
* conditions are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above
* copyright notice, this list of conditions and the following
* disclaimer in the documentation and/or other materials provided
* with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND
* CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES,
* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
* DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
* CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
* LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
* OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE,
* EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*
* ----------------------------------------------------------------------- */
/*
* bytesex.h - byte order helper functions
*
* In this function, be careful about getting X86_MEMORY versus
* LITTLE_ENDIAN correct: X86_MEMORY also means we are allowed to
* do unaligned memory references, and is probabilistic.
*/
#ifndef NASM_BYTEORD_H
#define NASM_BYTEORD_H
#include "compiler.h"
/*
* Some handy macros that will probably be of use in more than one
* output format: convert integers into little-endian byte packed
* format in memory.
*/
#define WRITECHAR(p,v) \
do { \
uint8_t *_wc_p = (uint8_t *)(p); \
*_wc_p++ = (v); \
(p) = (void *)_wc_p; \
} while (0)
#if X86_MEMORY
#define WRITESHORT(p,v) \
do { \
uint16_t *_ws_p = (uint16_t *)(p); \
*_ws_p++ = (v); \
(p) = (void *)_ws_p; \
} while (0)
#define WRITELONG(p,v) \
do { \
uint32_t *_wl_p = (uint32_t *)(p); \
*_wl_p++ = (v); \
(p) = (void *)_wl_p; \
} while (0)
#define WRITEDLONG(p,v) \
do { \
uint64_t *_wq_p = (uint64_t *)(p); \
*_wq_p++ = (v); \
(p) = (void *)_wq_p; \
} while (0)
#else /* !X86_MEMORY */
#define WRITESHORT(p,v) \
do { \
uint8_t *_ws_p = (uint8_t *)(p); \
const uint16_t _ws_v = (v); \
WRITECHAR(_ws_p, _ws_v); \
WRITECHAR(_ws_p, _ws_v >> 8); \
(p) = (void *)_ws_p; \
} while (0)
#define WRITELONG(p,v) \
do { \
uint8_t *_wl_p = (uint8_t *)(p); \
const uint32_t _wl_v = (v); \
WRITESHORT(_wl_p, _wl_v); \
WRITESHORT(_wl_p, _wl_v >> 16); \
(p) = (void *)_wl_p; \
} while (0)
#define WRITEDLONG(p,v) \
do { \
uint8_t *_wq_p = (uint8_t *)(p); \
const uint64_t _wq_v = (v); \
WRITELONG(_wq_p, _wq_v); \
WRITELONG(_wq_p, _wq_v >> 32); \
(p) = (void *)_wq_p; \
} while (0)
#endif /* X86_MEMORY */
/*
* Endian control functions which work on a single integer
*/
#ifdef WORDS_LITTLEENDIAN
#ifndef HAVE_CPU_TO_LE16
# define cpu_to_le16(v) ((uint16_t)(v))
#endif
#ifndef HAVE_CPU_TO_LE32
# define cpu_to_le32(v) ((uint32_t)(v))
#endif
#ifndef HAVE_CPU_TO_LE64
# define cpu_to_le64(v) ((uint64_t)(v))
#endif
#elif defined(WORDS_BIGENDIAN)
#ifndef HAVE_CPU_TO_LE16
static inline uint16_t cpu_to_le16(uint16_t v)
{
# ifdef HAVE___CPU_TO_LE16
return __cpu_to_le16(v);
# elif defined(HAVE_HTOLE16)
return htole16(v);
# elif defined(HAVE___BSWAP_16)
return __bswap_16(v);
# elif defined(HAVE___BUILTIN_BSWAP16)
return __builtin_bswap16(v);
# elif defined(HAVE__BYTESWAP_USHORT) && (USHRT_MAX == 0xffffU)
return _byteswap_ushort(v);
# else
return (v << 8) | (v >> 8);
# endif
}
#endif
#ifndef HAVE_CPU_TO_LE32
static inline uint32_t cpu_to_le32(uint32_t v)
{
# ifdef HAVE___CPU_TO_LE32
return __cpu_to_le32(v);
# elif defined(HAVE_HTOLE32)
return htole32(v);
# elif defined(HAVE___BSWAP_32)
return __bswap_32(v);
# elif defined(HAVE___BUILTIN_BSWAP32)
return __builtin_bswap32(v);
# elif defined(HAVE__BYTESWAP_ULONG) && (ULONG_MAX == 0xffffffffUL)
return _byteswap_ulong(v);
# else
v = ((v << 8) & 0xff00ff00 ) |
((v >> 8) & 0x00ff00ff);
return (v << 16) | (v >> 16);
# endif
}
#endif
#ifndef HAVE_CPU_TO_LE64
static inline uint64_t cpu_to_le64(uint64_t v)
{
# ifdef HAVE___CPU_TO_LE64
return __cpu_to_le64(v);
# elif defined(HAVE_HTOLE64)
return htole64(v);
# elif defined(HAVE___BSWAP_64)
return __bswap_64(v);
# elif defined(HAVE___BUILTIN_BSWAP64)
return __builtin_bswap64(v);
# elif defined(HAVE__BYTESWAP_UINT64)
return _byteswap_uint64(v);
# else
v = ((v << 8) & 0xff00ff00ff00ff00ull) |
((v >> 8) & 0x00ff00ff00ff00ffull);
v = ((v << 16) & 0xffff0000ffff0000ull) |
((v >> 16) & 0x0000ffff0000ffffull);
return (v << 32) | (v >> 32);
# endif
}
#endif
#else /* not WORDS_LITTLEENDIAN or WORDS_BIGENDIAN */
static inline uint16_t cpu_to_le16(uint16_t v)
{
union u16 {
uint16_t v;
uint8_t c[2];
} x;
uint8_t *cp = &x.c;
WRITESHORT(cp, v);
return x.v;
}
static inline uint32_t cpu_to_le32(uint32_t v)
{
union u32 {
uint32_t v;
uint8_t c[4];
} x;
uint8_t *cp = &x.c;
WRITELONG(cp, v);
return x.v;
}
static inline uint64_t cpu_to_le64(uint64_t v)
{
union u64 {
uint64_t v;
uint8_t c[8];
} x;
uint8_t *cp = &x.c;
WRITEDLONG(cp, v);
return x.v;
}
#endif
#define WRITEADDR(p,v,s) \
do { \
switch (is_constant(s) ? (s) : 0) { \
case 1: \
WRITECHAR(p,v); \
break; \
case 2: \
WRITESHORT(p,v); \
break; \
case 4: \
WRITELONG(p,v); \
break; \
case 8: \
WRITEDLONG(p,v); \
break; \
default: \
{ \
const uint64_t _wa_v = cpu_to_le64(v); \
const size_t _wa_s = (s); \
uint8_t * const _wa_p = (uint8_t *)(p); \
memcpy(_wa_p, &_wa_v, _wa_s); \
(p) = (void *)(_wa_p + _wa_s); \
} \
break; \
} \
} while (0)
#endif /* NASM_BYTESEX_H */