--- /dev/null Thu Jan 01 00:00:00 1970 +0000
+++ b/src/hotspot/os_cpu/solaris_x86/solaris_x86_64.il Tue Sep 12 19:03:39 2017 +0200
@@ -0,0 +1,135 @@
+//
+// Copyright (c) 2004, 2015, Oracle and/or its affiliates. All rights reserved.
+// DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
+//
+// This code is free software; you can redistribute it and/or modify it
+// under the terms of the GNU General Public License version 2 only, as
+// published by the Free Software Foundation.
+//
+// This code is distributed in the hope that it will be useful, but WITHOUT
+// ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+// FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
+// version 2 for more details (a copy is included in the LICENSE file that
+// accompanied this code).
+//
+// You should have received a copy of the GNU General Public License version
+// 2 along with this work; if not, write to the Free Software Foundation,
+// Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
+//
+// Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
+// or visit www.oracle.com if you need additional information or have any
+// questions.
+//
+//
+
+ // The argument size of each inline directive is ignored by the compiler
+ // and is set to the number of arguments as documentation.
+
+ // Get the raw thread ID from %gs:0
+ .inline _raw_thread_id,0
+ movq %fs:0, %rax
+ .end
+
+ // Get current sp
+ .inline _get_current_sp,0
+ .volatile
+ movq %rsp, %rax
+ .end
+
+ // Get current fp
+ .inline _get_current_fp,0
+ .volatile
+ movq %rbp, %rax
+ .end
+
+ // Support for os::rdtsc()
+ .inline _raw_rdtsc,0
+ rdtsc
+ salq $32, %rdx
+ orq %rdx, %rax
+ .end
+
+ // Support for jint Atomic::add(jint add_value, volatile jint* dest)
+ .inline _Atomic_add,2
+ movl %edi, %eax // save add_value for return
+ lock
+ xaddl %edi, (%rsi)
+ addl %edi, %eax
+ .end
+
+ // Support for jlong Atomic::add(jlong add_value, volatile jlong* dest)
+ .inline _Atomic_add_long,2
+ movq %rdi, %rax // save add_value for return
+ lock
+ xaddq %rdi, (%rsi)
+ addq %rdi, %rax
+ .end
+
+ // Support for jint Atomic::xchg(jint exchange_value, volatile jint* dest).
+ .inline _Atomic_xchg,2
+ xchgl (%rsi), %edi
+ movl %edi, %eax
+ .end
+
+ // Support for jlong Atomic::xchg(jlong exchange_value, volatile jlong* dest).
+ .inline _Atomic_xchg_long,2
+ xchgq (%rsi), %rdi
+ movq %rdi, %rax
+ .end
+
+ // Support for jbyte Atomic::cmpxchg(jbyte exchange_value,
+ // volatile jbyte *dest,
+ // jbyte compare_value)
+ .inline _Atomic_cmpxchg_byte,3
+ movb %dl, %al // compare_value
+ lock
+ cmpxchgb %dil, (%rsi)
+ .end
+
+ // Support for jint Atomic::cmpxchg(jint exchange_value,
+ // volatile jint *dest,
+ // jint compare_value)
+ .inline _Atomic_cmpxchg,3
+ movl %edx, %eax // compare_value
+ lock
+ cmpxchgl %edi, (%rsi)
+ .end
+
+ // Support for jlong Atomic::cmpxchg(jlong exchange_value,
+ // volatile jlong* dest,
+ // jlong compare_value)
+ .inline _Atomic_cmpxchg_long,3
+ movq %rdx, %rax // compare_value
+ lock
+ cmpxchgq %rdi, (%rsi)
+ .end
+
+ // Support for u2 Bytes::swap_u2(u2 x)
+ .inline _raw_swap_u2,1
+ movw %di, %ax
+ rorw $8, %ax
+ .end
+
+ // Support for u4 Bytes::swap_u4(u4 x)
+ .inline _raw_swap_u4,1
+ movl %edi, %eax
+ bswapl %eax
+ .end
+
+ // Support for u8 Bytes::swap_u8(u8 x)
+ .inline _raw_swap_u8,1
+ movq %rdi, %rax
+ bswapq %rax
+ .end
+
+ // Support for void Prefetch::read
+ .inline _Prefetch_read,2
+ prefetcht0 (%rdi, %rsi, 1)
+ .end
+
+ // Support for void Prefetch::write
+ // We use prefetcht0 because em64t doesn't support prefetchw.
+ // prefetchw is a 3dnow instruction.
+ .inline _Prefetch_write,2
+ prefetcht0 (%rdi, %rsi, 1)
+ .end