--- /dev/null Thu Jan 01 00:00:00 1970 +0000
+++ b/src/hotspot/cpu/s390/c1_MacroAssembler_s390.cpp Tue Sep 12 19:03:39 2017 +0200
@@ -0,0 +1,380 @@
+/*
+ * Copyright (c) 2016, Oracle and/or its affiliates. All rights reserved.
+ * Copyright (c) 2016 SAP SE. All rights reserved.
+ * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
+ *
+ * This code is free software; you can redistribute it and/or modify it
+ * under the terms of the GNU General Public License version 2 only, as
+ * published by the Free Software Foundation.
+ *
+ * This code is distributed in the hope that it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
+ * version 2 for more details (a copy is included in the LICENSE file that
+ * accompanied this code).
+ *
+ * You should have received a copy of the GNU General Public License version
+ * 2 along with this work; if not, write to the Free Software Foundation,
+ * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
+ * or visit www.oracle.com if you need additional information or have any
+ * questions.
+ *
+ */
+
+#include "precompiled.hpp"
+#include "c1/c1_MacroAssembler.hpp"
+#include "c1/c1_Runtime1.hpp"
+#include "classfile/systemDictionary.hpp"
+#include "gc/shared/collectedHeap.hpp"
+#include "interpreter/interpreter.hpp"
+#include "oops/arrayOop.hpp"
+#include "oops/markOop.hpp"
+#include "runtime/basicLock.hpp"
+#include "runtime/biasedLocking.hpp"
+#include "runtime/os.hpp"
+#include "runtime/sharedRuntime.hpp"
+#include "runtime/stubRoutines.hpp"
+
+void C1_MacroAssembler::inline_cache_check(Register receiver, Register iCache) {
+ Label ic_miss, ic_hit;
+ verify_oop(receiver);
+ int klass_offset = oopDesc::klass_offset_in_bytes();
+
+ if (!ImplicitNullChecks || MacroAssembler::needs_explicit_null_check(klass_offset)) {
+ if (VM_Version::has_CompareBranch()) {
+ z_cgij(receiver, 0, Assembler::bcondEqual, ic_miss);
+ } else {
+ z_ltgr(receiver, receiver);
+ z_bre(ic_miss);
+ }
+ }
+
+ compare_klass_ptr(iCache, klass_offset, receiver, false);
+ z_bre(ic_hit);
+
+ // If icache check fails, then jump to runtime routine.
+ // Note: RECEIVER must still contain the receiver!
+ load_const_optimized(Z_R1_scratch, AddressLiteral(SharedRuntime::get_ic_miss_stub()));
+ z_br(Z_R1_scratch);
+ align(CodeEntryAlignment);
+ bind(ic_hit);
+}
+
+void C1_MacroAssembler::explicit_null_check(Register base) {
+ ShouldNotCallThis(); // unused
+}
+
+void C1_MacroAssembler::build_frame(int frame_size_in_bytes, int bang_size_in_bytes) {
+ assert(bang_size_in_bytes >= frame_size_in_bytes, "stack bang size incorrect");
+ generate_stack_overflow_check(bang_size_in_bytes);
+ save_return_pc();
+ push_frame(frame_size_in_bytes);
+}
+
+void C1_MacroAssembler::unverified_entry(Register receiver, Register ic_klass) {
+ ShouldNotCallThis(); // unused
+}
+
+void C1_MacroAssembler::verified_entry() {
+ if (C1Breakpoint) z_illtrap(0xC1);
+}
+
+void C1_MacroAssembler::lock_object(Register hdr, Register obj, Register disp_hdr, Label& slow_case) {
+ const int hdr_offset = oopDesc::mark_offset_in_bytes();
+ assert_different_registers(hdr, obj, disp_hdr);
+ NearLabel done;
+
+ verify_oop(obj);
+
+ // Load object header.
+ z_lg(hdr, Address(obj, hdr_offset));
+
+ // Save object being locked into the BasicObjectLock...
+ z_stg(obj, Address(disp_hdr, BasicObjectLock::obj_offset_in_bytes()));
+
+ if (UseBiasedLocking) {
+ biased_locking_enter(obj, hdr, Z_R1_scratch, Z_R0_scratch, done, &slow_case);
+ }
+
+ // and mark it as unlocked.
+ z_oill(hdr, markOopDesc::unlocked_value);
+ // Save unlocked object header into the displaced header location on the stack.
+ z_stg(hdr, Address(disp_hdr, (intptr_t)0));
+ // Test if object header is still the same (i.e. unlocked), and if so, store the
+ // displaced header address in the object header. If it is not the same, get the
+ // object header instead.
+ z_csg(hdr, disp_hdr, hdr_offset, obj);
+ // If the object header was the same, we're done.
+ if (PrintBiasedLockingStatistics) {
+ Unimplemented();
+#if 0
+ cond_inc32(Assembler::equal,
+ ExternalAddress((address)BiasedLocking::fast_path_entry_count_addr()));
+#endif
+ }
+ branch_optimized(Assembler::bcondEqual, done);
+ // If the object header was not the same, it is now in the hdr register.
+ // => Test if it is a stack pointer into the same stack (recursive locking), i.e.:
+ //
+ // 1) (hdr & markOopDesc::lock_mask_in_place) == 0
+ // 2) rsp <= hdr
+ // 3) hdr <= rsp + page_size
+ //
+ // These 3 tests can be done by evaluating the following expression:
+ //
+ // (hdr - Z_SP) & (~(page_size-1) | markOopDesc::lock_mask_in_place)
+ //
+ // assuming both the stack pointer and page_size have their least
+ // significant 2 bits cleared and page_size is a power of 2
+ z_sgr(hdr, Z_SP);
+
+ load_const_optimized(Z_R0_scratch, (~(os::vm_page_size()-1) | markOopDesc::lock_mask_in_place));
+ z_ngr(hdr, Z_R0_scratch); // AND sets CC (result eq/ne 0).
+ // For recursive locking, the result is zero. => Save it in the displaced header
+ // location (NULL in the displaced hdr location indicates recursive locking).
+ z_stg(hdr, Address(disp_hdr, (intptr_t)0));
+ // Otherwise we don't care about the result and handle locking via runtime call.
+ branch_optimized(Assembler::bcondNotZero, slow_case);
+ // done
+ bind(done);
+}
+
+void C1_MacroAssembler::unlock_object(Register hdr, Register obj, Register disp_hdr, Label& slow_case) {
+ const int aligned_mask = BytesPerWord -1;
+ const int hdr_offset = oopDesc::mark_offset_in_bytes();
+ assert_different_registers(hdr, obj, disp_hdr);
+ NearLabel done;
+
+ if (UseBiasedLocking) {
+ // Load object.
+ z_lg(obj, Address(disp_hdr, BasicObjectLock::obj_offset_in_bytes()));
+ biased_locking_exit(obj, hdr, done);
+ }
+
+ // Load displaced header.
+ z_ltg(hdr, Address(disp_hdr, (intptr_t)0));
+ // If the loaded hdr is NULL we had recursive locking, and we are done.
+ z_bre(done);
+ if (!UseBiasedLocking) {
+ // Load object.
+ z_lg(obj, Address(disp_hdr, BasicObjectLock::obj_offset_in_bytes()));
+ }
+ verify_oop(obj);
+ // Test if object header is pointing to the displaced header, and if so, restore
+ // the displaced header in the object. If the object header is not pointing to
+ // the displaced header, get the object header instead.
+ z_csg(disp_hdr, hdr, hdr_offset, obj);
+ // If the object header was not pointing to the displaced header,
+ // we do unlocking via runtime call.
+ branch_optimized(Assembler::bcondNotEqual, slow_case);
+ // done
+ bind(done);
+}
+
+void C1_MacroAssembler::try_allocate(
+ Register obj, // result: Pointer to object after successful allocation.
+ Register var_size_in_bytes, // Object size in bytes if unknown at compile time; invalid otherwise.
+ int con_size_in_bytes, // Object size in bytes if known at compile time.
+ Register t1, // Temp register: Must be global register for incr_allocated_bytes.
+ Label& slow_case // Continuation point if fast allocation fails.
+) {
+ if (UseTLAB) {
+ tlab_allocate(obj, var_size_in_bytes, con_size_in_bytes, t1, slow_case);
+ } else {
+ // Allocation in shared Eden not implemented, because sapjvm allocation trace does not allow it.
+ z_brul(slow_case);
+ }
+}
+
+void C1_MacroAssembler::initialize_header(Register obj, Register klass, Register len, Register Rzero, Register t1) {
+ assert_different_registers(obj, klass, len, t1, Rzero);
+ if (UseBiasedLocking && !len->is_valid()) {
+ assert_different_registers(obj, klass, len, t1);
+ z_lg(t1, Address(klass, Klass::prototype_header_offset()));
+ } else {
+ // This assumes that all prototype bits fit in an int32_t.
+ load_const_optimized(t1, (intx)markOopDesc::prototype());
+ }
+ z_stg(t1, Address(obj, oopDesc::mark_offset_in_bytes()));
+
+ if (len->is_valid()) {
+ // Length will be in the klass gap, if one exists.
+ z_st(len, Address(obj, arrayOopDesc::length_offset_in_bytes()));
+ } else if (UseCompressedClassPointers) {
+ store_klass_gap(Rzero, obj); // Zero klass gap for compressed oops.
+ }
+ store_klass(klass, obj, t1);
+}
+
+void C1_MacroAssembler::initialize_body(Register objectFields, Register len_in_bytes, Register Rzero) {
+ Label done;
+ assert_different_registers(objectFields, len_in_bytes, Rzero);
+
+ // Initialize object fields.
+ // See documentation for MVCLE instruction!!!
+ assert(objectFields->encoding()%2==0, "objectFields must be an even register");
+ assert(len_in_bytes->encoding() == (objectFields->encoding()+1), "objectFields and len_in_bytes must be a register pair");
+ assert(Rzero->encoding()%2==1, "Rzero must be an odd register");
+
+ // Use Rzero as src length, then mvcle will copy nothing
+ // and fill the object with the padding value 0.
+ move_long_ext(objectFields, as_Register(Rzero->encoding()-1), 0);
+ bind(done);
+}
+
+void C1_MacroAssembler::allocate_object(
+ Register obj, // Result: pointer to object after successful allocation.
+ Register t1, // temp register
+ Register t2, // temp register: Must be a global register for try_allocate.
+ int hdr_size, // object header size in words
+ int obj_size, // object size in words
+ Register klass, // object klass
+ Label& slow_case // Continuation point if fast allocation fails.
+) {
+ assert_different_registers(obj, t1, t2, klass);
+
+ // Allocate space and initialize header.
+ try_allocate(obj, noreg, obj_size * wordSize, t1, slow_case);
+
+ initialize_object(obj, klass, noreg, obj_size * HeapWordSize, t1, t2);
+}
+
+void C1_MacroAssembler::initialize_object(
+ Register obj, // result: Pointer to object after successful allocation.
+ Register klass, // object klass
+ Register var_size_in_bytes, // Object size in bytes if unknown at compile time; invalid otherwise.
+ int con_size_in_bytes, // Object size in bytes if known at compile time.
+ Register t1, // temp register
+ Register t2 // temp register
+ ) {
+ assert((con_size_in_bytes & MinObjAlignmentInBytesMask) == 0,
+ "con_size_in_bytes is not multiple of alignment");
+ assert(var_size_in_bytes == noreg, "not implemented");
+ const int hdr_size_in_bytes = instanceOopDesc::header_size() * HeapWordSize;
+
+ const Register Rzero = t2;
+
+ z_xgr(Rzero, Rzero);
+ initialize_header(obj, klass, noreg, Rzero, t1);
+
+ // Clear rest of allocated space.
+ const int threshold = 4 * BytesPerWord;
+ if (con_size_in_bytes <= threshold) {
+ // Use explicit null stores.
+ // code size = 6*n bytes (n = number of fields to clear)
+ for (int i = hdr_size_in_bytes; i < con_size_in_bytes; i += BytesPerWord)
+ z_stg(Rzero, Address(obj, i));
+ } else {
+ // Code size generated by initialize_body() is 16.
+ Register object_fields = Z_R0_scratch;
+ Register len_in_bytes = Z_R1_scratch;
+ z_la(object_fields, hdr_size_in_bytes, obj);
+ load_const_optimized(len_in_bytes, con_size_in_bytes - hdr_size_in_bytes);
+ initialize_body(object_fields, len_in_bytes, Rzero);
+ }
+
+ // Dtrace support is unimplemented.
+ // if (CURRENT_ENV->dtrace_alloc_probes()) {
+ // assert(obj == rax, "must be");
+ // call(RuntimeAddress(Runtime1::entry_for (Runtime1::dtrace_object_alloc_id)));
+ // }
+
+ verify_oop(obj);
+}
+
+void C1_MacroAssembler::allocate_array(
+ Register obj, // result: Pointer to array after successful allocation.
+ Register len, // array length
+ Register t1, // temp register
+ Register t2, // temp register
+ int hdr_size, // object header size in words
+ int elt_size, // element size in bytes
+ Register klass, // object klass
+ Label& slow_case // Continuation point if fast allocation fails.
+) {
+ assert_different_registers(obj, len, t1, t2, klass);
+
+ // Determine alignment mask.
+ assert(!(BytesPerWord & 1), "must be a multiple of 2 for masking code to work");
+
+ // Check for negative or excessive length.
+ compareU64_and_branch(len, (int32_t)max_array_allocation_length, bcondHigh, slow_case);
+
+ // Compute array size.
+ // Note: If 0 <= len <= max_length, len*elt_size + header + alignment is
+ // smaller or equal to the largest integer. Also, since top is always
+ // aligned, we can do the alignment here instead of at the end address
+ // computation.
+ const Register arr_size = t2;
+ switch (elt_size) {
+ case 1: lgr_if_needed(arr_size, len); break;
+ case 2: z_sllg(arr_size, len, 1); break;
+ case 4: z_sllg(arr_size, len, 2); break;
+ case 8: z_sllg(arr_size, len, 3); break;
+ default: ShouldNotReachHere();
+ }
+ add2reg(arr_size, hdr_size * wordSize + MinObjAlignmentInBytesMask); // Add space for header & alignment.
+ z_nill(arr_size, (~MinObjAlignmentInBytesMask) & 0xffff); // Align array size.
+
+ try_allocate(obj, arr_size, 0, t1, slow_case);
+
+ initialize_header(obj, klass, len, noreg, t1);
+
+ // Clear rest of allocated space.
+ Label done;
+ Register object_fields = t1;
+ Register Rzero = Z_R1_scratch;
+ z_aghi(arr_size, -(hdr_size * BytesPerWord));
+ z_bre(done); // Jump if size of fields is zero.
+ z_la(object_fields, hdr_size * BytesPerWord, obj);
+ z_xgr(Rzero, Rzero);
+ initialize_body(object_fields, arr_size, Rzero);
+ bind(done);
+
+ // Dtrace support is unimplemented.
+ // if (CURRENT_ENV->dtrace_alloc_probes()) {
+ // assert(obj == rax, "must be");
+ // call(RuntimeAddress(Runtime1::entry_for (Runtime1::dtrace_object_alloc_id)));
+ // }
+
+ verify_oop(obj);
+}
+
+
+#ifndef PRODUCT
+
+void C1_MacroAssembler::verify_stack_oop(int stack_offset) {
+ Unimplemented();
+ // if (!VerifyOops) return;
+ // verify_oop_addr(Address(SP, stack_offset + STACK_BIAS));
+}
+
+void C1_MacroAssembler::verify_not_null_oop(Register r) {
+ if (!VerifyOops) return;
+ NearLabel not_null;
+ compareU64_and_branch(r, (intptr_t)0, bcondNotEqual, not_null);
+ stop("non-null oop required");
+ bind(not_null);
+ verify_oop(r);
+}
+
+void C1_MacroAssembler::invalidate_registers(Register preserve1,
+ Register preserve2,
+ Register preserve3) {
+ Register dead_value = noreg;
+ for (int i = 0; i < FrameMap::nof_cpu_regs; i++) {
+ Register r = as_Register(i);
+ if (r != preserve1 && r != preserve2 && r != preserve3 && r != Z_SP && r != Z_thread) {
+ if (dead_value == noreg) {
+ load_const_optimized(r, 0xc1dead);
+ dead_value = r;
+ } else {
+ z_lgr(r, dead_value);
+ }
+ }
+ }
+}
+
+#endif // !PRODUCT