1 /* 2 * Copyright (c) 1999, 2010, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #ifndef OS_CPU_BSD_X86_VM_BYTES_BSD_X86_INLINE_HPP 26 #define OS_CPU_BSD_X86_VM_BYTES_BSD_X86_INLINE_HPP 27 28 #ifndef _ALLBSD_SOURCE 29 #include <byteswap.h> 30 #endif 31 32 #ifdef __APPLE__ 33 #include <libkern/OSByteOrder.h> 34 #endif 35 36 #if defined(AMD64) 37 # if defined(__APPLE__) 38 # define bswap_16(x) OSSwapInt16(x) 39 # define bswap_32(x) OSSwapInt32(x) 40 # define bswap_64(x) OSSwapInt64(x) 41 # elif defined(__OpenBSD__) 42 # define bswap_16(x) swap16(x) 43 # define bswap_32(x) swap32(x) 44 # define bswap_64(x) swap64(x) 45 # elif defined(__NetBSD__) 46 # define bswap_16(x) bswap16(x) 47 # define bswap_32(x) bswap32(x) 48 # define bswap_64(x) bswap64(x) 49 # else 50 # define bswap_16(x) __bswap16(x) 51 # define bswap_32(x) __bswap32(x) 52 # define bswap_64(x) __bswap64(x) 53 # endif 54 #endif 55 56 // Efficient swapping of data bytes from Java byte 57 // ordering to native byte ordering and vice versa. 58 inline u2 Bytes::swap_u2(u2 x) { 59 #ifdef AMD64 60 return bswap_16(x); 61 #else 62 u2 ret; 63 __asm__ __volatile__ ( 64 "movw %0, %%ax;" 65 "xchg %%al, %%ah;" 66 "movw %%ax, %0" 67 :"=r" (ret) // output : register 0 => ret 68 :"0" (x) // input : x => register 0 69 :"ax", "0" // clobbered registers 70 ); 71 return ret; 72 #endif // AMD64 73 } 74 75 inline u4 Bytes::swap_u4(u4 x) { 76 #ifdef AMD64 77 return bswap_32(x); 78 #else 79 u4 ret; 80 __asm__ __volatile__ ( 81 "bswap %0" 82 :"=r" (ret) // output : register 0 => ret 83 :"0" (x) // input : x => register 0 84 :"0" // clobbered register 85 ); 86 return ret; 87 #endif // AMD64 88 } 89 90 #ifdef AMD64 91 inline u8 Bytes::swap_u8(u8 x) { 92 #ifdef SPARC_WORKS 93 // workaround for SunStudio12 CR6615391 94 __asm__ __volatile__ ( 95 "bswapq %0" 96 :"=r" (x) // output : register 0 => x 97 :"0" (x) // input : x => register 0 98 :"0" // clobbered register 99 ); 100 return x; 101 #else 102 return bswap_64(x); 103 #endif 104 } 105 #else 106 // Helper function for swap_u8 107 inline u8 Bytes::swap_u8_base(u4 x, u4 y) { 108 return (((u8)swap_u4(x))<<32) | swap_u4(y); 109 } 110 111 inline u8 Bytes::swap_u8(u8 x) { 112 return swap_u8_base(*(u4*)&x, *(((u4*)&x)+1)); 113 } 114 #endif // !AMD64 115 116 #endif // OS_CPU_BSD_X86_VM_BYTES_BSD_X86_INLINE_HPP