1 //
   2 // Copyright (c) 2004, 2012, Oracle and/or its affiliates. All rights reserved.
   3 // DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   4 //
   5 // This code is free software; you can redistribute it and/or modify it
   6 // under the terms of the GNU General Public License version 2 only, as
   7 // published by the Free Software Foundation.
   8 //
   9 // This code is distributed in the hope that it will be useful, but WITHOUT
  10 // ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  11 // FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  12 // version 2 for more details (a copy is included in the LICENSE file that
  13 // accompanied this code).
  14 //
  15 // You should have received a copy of the GNU General Public License version
  16 // 2 along with this work; if not, write to the Free Software Foundation,
  17 // Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  18 //
  19 // Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
  20 // or visit www.oracle.com if you need additional information or have any
  21 // questions.
  22 //
  23 //
  24 
  25   // The argument size of each inline directive is ignored by the compiler
  26   // and is set to the number of arguments as documentation.
  27 
  28   // Get the raw thread ID from %gs:0
  29       .inline _raw_thread_id,0
  30       movq     %fs:0, %rax
  31       .end
  32 
  33   // Get current sp
  34       .inline _get_current_sp,0
  35       .volatile
  36       movq     %rsp, %rax
  37       .end
  38 
  39   // Get current fp
  40       .inline _get_current_fp,0
  41       .volatile
  42       movq     %rbp, %rax
  43       .end
  44 
  45   // Support for os::rdtsc()
  46       .inline _raw_rdtsc,0
  47       rdtsc
  48       salq     $32, %rdx
  49       orq      %rdx, %rax
  50       .end
  51 
  52   // Support for jint Atomic::add(jint add_value, volatile jint* dest)
  53       .inline _Atomic_add,2
  54       movl     %edi, %eax      // save add_value for return
  55       lock
  56       xaddl    %edi, (%rsi)
  57       addl     %edi, %eax
  58       .end
  59 
  60   // Support for jlong Atomic::add(jlong add_value, volatile jlong* dest)
  61       .inline _Atomic_add_long,2
  62       movq     %rdi, %rax      // save add_value for return
  63       lock
  64       xaddq    %rdi, (%rsi)
  65       addq     %rdi, %rax
  66       .end
  67 
  68   // Support for jint Atomic::xchg(jint exchange_value, volatile jint* dest).
  69       .inline _Atomic_xchg,2
  70       xchgl    (%rsi), %edi
  71       movl     %edi, %eax
  72       .end
  73 
  74   // Support for jlong Atomic::xchg(jlong exchange_value, volatile jlong* dest).
  75       .inline _Atomic_xchg_long,2
  76       xchgq    (%rsi), %rdi
  77       movq     %rdi, %rax
  78       .end
  79 
  80   // Support for jbyte Atomic::cmpxchg(jbyte exchange_value,
  81   //                                   volatile jbyte *dest,
  82   //                                   jbyte compare_value)
  83       .inline _Atomic_cmpxchg_byte,3
  84       movb     %dl, %al      // compare_value
  85       lock
  86       cmpxchgb %dil, (%rsi)
  87       .end
  88 
  89   // Support for jint Atomic::cmpxchg(jint exchange_value,
  90   //                                  volatile jint *dest,
  91   //                                  jint compare_value)
  92       .inline _Atomic_cmpxchg,3
  93       movl     %edx, %eax      // compare_value
  94       lock
  95       cmpxchgl %edi, (%rsi)
  96       .end
  97 
  98   // Support for jlong Atomic::cmpxchg(jlong exchange_value,
  99   //                                   volatile jlong* dest,
 100   //                                   jlong compare_value)
 101       .inline _Atomic_cmpxchg_long,3
 102       movq     %rdx, %rax      // compare_value
 103       lock
 104       cmpxchgq %rdi, (%rsi)
 105       .end
 106 
 107   // Support for u2 Bytes::swap_u2(u2 x)
 108       .inline _raw_swap_u2,1
 109       movw     %di, %ax
 110       rorw     $8, %ax
 111       .end
 112 
 113   // Support for u4 Bytes::swap_u4(u4 x)
 114       .inline _raw_swap_u4,1
 115       movl     %edi, %eax
 116       bswapl   %eax
 117       .end
 118 
 119   // Support for u8 Bytes::swap_u8(u8 x)
 120       .inline _raw_swap_u8,1
 121       movq     %rdi, %rax
 122       bswapq   %rax
 123       .end
 124 
 125   // Support for void Prefetch::read
 126       .inline _Prefetch_read,2
 127       prefetcht0 (%rdi, %rsi, 1)
 128       .end
 129 
 130   // Support for void Prefetch::write
 131   // We use prefetcht0 because em64t doesn't support prefetchw.
 132   // prefetchw is a 3dnow instruction.
 133       .inline _Prefetch_write,2
 134       prefetcht0 (%rdi, %rsi, 1)
 135       .end