1 /*
   2  * Copyright (c) 1999, 2019, Oracle and/or its affiliates. All rights reserved.
   3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   4  *
   5  * This code is free software; you can redistribute it and/or modify it
   6  * under the terms of the GNU General Public License version 2 only, as
   7  * published by the Free Software Foundation.
   8  *
   9  * This code is distributed in the hope that it will be useful, but WITHOUT
  10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  12  * version 2 for more details (a copy is included in the LICENSE file that
  13  * accompanied this code).
  14  *
  15  * You should have received a copy of the GNU General Public License version
  16  * 2 along with this work; if not, write to the Free Software Foundation,
  17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  18  *
  19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
  20  * or visit www.oracle.com if you need additional information or have any
  21  * questions.
  22  *
  23  */
  24 
  25 #ifndef OS_CPU_WINDOWS_X86_OS_WINDOWS_X86_HPP
  26 #define OS_CPU_WINDOWS_X86_OS_WINDOWS_X86_HPP
  27 
  28   //
  29   // NOTE: we are back in class os here, not win32
  30   //
  31 #ifdef AMD64
  32   static int32_t   (*atomic_xchg_func)          (int32_t, volatile int32_t*);
  33   static int64_t   (*atomic_xchg_long_func)     (int64_t, volatile int64_t*);
  34 
  35   static int32_t   (*atomic_cmpxchg_func)       (int32_t,  volatile int32_t*, int32_t);
  36   static int8_t    (*atomic_cmpxchg_byte_func)  (int8_t,   volatile int8_t*,  int8_t);
  37   static int64_t   (*atomic_cmpxchg_long_func)  (int64_t,  volatile int64_t*, int64_t);
  38 
  39   static int32_t   (*atomic_add_func)           (int32_t,  volatile int32_t*);
  40   static int64_t   (*atomic_add_long_func)      (int64_t,  volatile int64_t*);
  41 
  42   static int32_t   atomic_xchg_bootstrap        (int32_t,  volatile int32_t*);
  43   static int64_t   atomic_xchg_long_bootstrap   (int64_t,  volatile int64_t*);
  44 
  45   static int32_t   atomic_cmpxchg_bootstrap     (int32_t,  volatile int32_t*, int32_t);
  46   static int8_t    atomic_cmpxchg_byte_bootstrap(int8_t,   volatile int8_t*,  int8_t);
  47 #else
  48 
  49   static int64_t (*atomic_cmpxchg_long_func)  (int64_t, volatile int64_t*, int64_t);
  50 
  51 #endif // AMD64
  52 
  53   static int64_t atomic_cmpxchg_long_bootstrap(int64_t, volatile int64_t*, int64_t);
  54 
  55 #ifdef AMD64
  56   static int32_t  atomic_add_bootstrap         (int32_t,  volatile int32_t*);
  57   static int64_t  atomic_add_long_bootstrap    (int64_t,  volatile int64_t*);
  58 #endif // AMD64
  59 
  60   static void setup_fpu();
  61   static bool supports_sse() { return true; }
  62   static juint cpu_microcode_revision();
  63 
  64   static jlong rdtsc();
  65 
  66   static bool      register_code_area(char *low, char *high);
  67 
  68 #ifdef AMD64
  69 #define PLATFORM_PRINT_NATIVE_STACK 1
  70 static bool platform_print_native_stack(outputStream* st, const void* context,
  71                                         char *buf, int buf_size);
  72 #endif
  73 
  74 #endif // OS_CPU_WINDOWS_X86_OS_WINDOWS_X86_HPP