1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or https://opensource.org/licenses/CDDL-1.0. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright 2007 Sun Microsystems, Inc. All rights reserved. 24 * Use is subject to license terms. 25 */ 26 27 /* Copyright (c) 1983, 1984, 1985, 1986, 1987, 1988, 1989 AT&T */ 28 /* All Rights Reserved */ 29 30 /* 31 * University Copyright- Copyright (c) 1982, 1986, 1988 32 * The Regents of the University of California 33 * All Rights Reserved 34 * 35 * University Acknowledgment- Portions of this document are derived from 36 * software developed by the University of California, Berkeley, and its 37 * contributors. 38 */ 39 40 #ifndef _SYS_BYTEORDER_H 41 #define _SYS_BYTEORDER_H 42 43 #include <sys/endian.h> 44 #include <netinet/in.h> 45 #include <sys/isa_defs.h> 46 #include <inttypes.h> 47 48 #if defined(__GNUC__) && defined(_ASM_INLINES) && \ 49 (defined(__i386) || defined(__amd64)) 50 #include <asm/byteorder.h> 51 #endif 52 53 #ifdef __cplusplus 54 extern "C" { 55 #endif 56 57 /* 58 * macros for conversion between host and (internet) network byte order 59 */ 60 #if !defined(_XPG4_2) || defined(__EXTENSIONS__) 61 62 #ifdef __COVERITY__ 63 /* 64 * Coverity's taint warnings from byteswapping are false positives for us. 65 * Suppress them by hiding byteswapping from Coverity. 66 */ 67 #define BSWAP_8(x) ((x) & 0xff) 68 #define BSWAP_16(x) ((x) & 0xffff) 69 #define BSWAP_32(x) ((x) & 0xffffffff) 70 #define BSWAP_64(x) (x) 71 72 #else /* __COVERITY__ */ 73 74 /* 75 * Macros to reverse byte order 76 */ 77 #define BSWAP_8(x) ((x) & 0xff) 78 #define BSWAP_16(x) ((BSWAP_8(x) << 8) | BSWAP_8((x) >> 8)) 79 #define BSWAP_32(x) ((BSWAP_16(x) << 16) | BSWAP_16((x) >> 16)) 80 #define BSWAP_64(x) ((BSWAP_32(x) << 32) | BSWAP_32((x) >> 32)) 81 82 #endif /* __COVERITY__ */ 83 84 #define BMASK_8(x) ((x) & 0xff) 85 #define BMASK_16(x) ((x) & 0xffff) 86 #define BMASK_32(x) ((x) & 0xffffffff) 87 #define BMASK_64(x) (x) 88 89 /* 90 * Macros to convert from a specific byte order to/from native byte order 91 */ 92 #ifdef _ZFS_BIG_ENDIAN 93 #define BE_8(x) BMASK_8(x) 94 #define BE_16(x) BMASK_16(x) 95 #define BE_32(x) BMASK_32(x) 96 #define BE_64(x) BMASK_64(x) 97 #define LE_8(x) BSWAP_8(x) 98 #define LE_16(x) BSWAP_16(x) 99 #define LE_32(x) BSWAP_32(x) 100 #define LE_64(x) BSWAP_64(x) 101 #else 102 #define LE_8(x) BMASK_8(x) 103 #define LE_16(x) BMASK_16(x) 104 #define LE_32(x) BMASK_32(x) 105 #define LE_64(x) BMASK_64(x) 106 #define BE_8(x) BSWAP_8(x) 107 #define BE_16(x) BSWAP_16(x) 108 #define BE_32(x) BSWAP_32(x) 109 #define BE_64(x) BSWAP_64(x) 110 #endif 111 112 #ifdef _ZFS_BIG_ENDIAN 113 static __inline__ uint64_t 114 htonll(uint64_t n) 115 { 116 return (n); 117 } 118 119 static __inline__ uint64_t 120 ntohll(uint64_t n) 121 { 122 return (n); 123 } 124 #else 125 static __inline__ uint64_t 126 htonll(uint64_t n) 127 { 128 return ((((uint64_t)htonl(n)) << 32) + htonl(n >> 32)); 129 } 130 131 static __inline__ uint64_t 132 ntohll(uint64_t n) 133 { 134 return ((((uint64_t)ntohl(n)) << 32) + ntohl(n >> 32)); 135 } 136 #endif 137 138 /* 139 * Macros to read unaligned values from a specific byte order to 140 * native byte order 141 */ 142 143 #define BE_IN8(xa) \ 144 *((uint8_t *)(xa)) 145 146 #define BE_IN16(xa) \ 147 (((uint16_t)BE_IN8(xa) << 8) | BE_IN8((uint8_t *)(xa)+1)) 148 149 #define BE_IN32(xa) \ 150 (((uint32_t)BE_IN16(xa) << 16) | BE_IN16((uint8_t *)(xa)+2)) 151 152 #define BE_IN64(xa) \ 153 (((uint64_t)BE_IN32(xa) << 32) | BE_IN32((uint8_t *)(xa)+4)) 154 155 #define LE_IN8(xa) \ 156 *((uint8_t *)(xa)) 157 158 #define LE_IN16(xa) \ 159 (((uint16_t)LE_IN8((uint8_t *)(xa) + 1) << 8) | LE_IN8(xa)) 160 161 #define LE_IN32(xa) \ 162 (((uint32_t)LE_IN16((uint8_t *)(xa) + 2) << 16) | LE_IN16(xa)) 163 164 #define LE_IN64(xa) \ 165 (((uint64_t)LE_IN32((uint8_t *)(xa) + 4) << 32) | LE_IN32(xa)) 166 167 /* 168 * Macros to write unaligned values from native byte order to a specific byte 169 * order. 170 */ 171 172 #define BE_OUT8(xa, yv) *((uint8_t *)(xa)) = (uint8_t)(yv); 173 174 #define BE_OUT16(xa, yv) \ 175 BE_OUT8((uint8_t *)(xa) + 1, yv); \ 176 BE_OUT8((uint8_t *)(xa), (yv) >> 8); 177 178 #define BE_OUT32(xa, yv) \ 179 BE_OUT16((uint8_t *)(xa) + 2, yv); \ 180 BE_OUT16((uint8_t *)(xa), (yv) >> 16); 181 182 #define BE_OUT64(xa, yv) \ 183 BE_OUT32((uint8_t *)(xa) + 4, yv); \ 184 BE_OUT32((uint8_t *)(xa), (yv) >> 32); 185 186 #define LE_OUT8(xa, yv) *((uint8_t *)(xa)) = (uint8_t)(yv); 187 188 #define LE_OUT16(xa, yv) \ 189 LE_OUT8((uint8_t *)(xa), yv); \ 190 LE_OUT8((uint8_t *)(xa) + 1, (yv) >> 8); 191 192 #define LE_OUT32(xa, yv) \ 193 LE_OUT16((uint8_t *)(xa), yv); \ 194 LE_OUT16((uint8_t *)(xa) + 2, (yv) >> 16); 195 196 #define LE_OUT64(xa, yv) \ 197 LE_OUT32((uint8_t *)(xa), yv); \ 198 LE_OUT32((uint8_t *)(xa) + 4, (yv) >> 32); 199 200 #endif /* !defined(_XPG4_2) || defined(__EXTENSIONS__) */ 201 202 #ifdef __cplusplus 203 } 204 #endif 205 206 #endif /* _SYS_BYTEORDER_H */ 207