35085
|
1 |
/*
|
|
2 |
* Copyright (c) 1999, 2015, Oracle and/or its affiliates. All rights reserved.
|
35594
|
3 |
* Copyright (c) 2012, 2015 SAP SE. All rights reserved.
|
35085
|
4 |
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
5 |
*
|
|
6 |
* This code is free software; you can redistribute it and/or modify it
|
|
7 |
* under the terms of the GNU General Public License version 2 only, as
|
|
8 |
* published by the Free Software Foundation.
|
|
9 |
*
|
|
10 |
* This code is distributed in the hope that it will be useful, but WITHOUT
|
|
11 |
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
12 |
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
13 |
* version 2 for more details (a copy is included in the LICENSE file that
|
|
14 |
* accompanied this code).
|
|
15 |
*
|
|
16 |
* You should have received a copy of the GNU General Public License version
|
|
17 |
* 2 along with this work; if not, write to the Free Software Foundation,
|
|
18 |
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
19 |
*
|
|
20 |
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
21 |
* or visit www.oracle.com if you need additional information or have any
|
|
22 |
* questions.
|
|
23 |
*
|
|
24 |
*/
|
|
25 |
|
|
26 |
#include "precompiled.hpp"
|
|
27 |
#include "c1/c1_MacroAssembler.hpp"
|
|
28 |
#include "c1/c1_Runtime1.hpp"
|
|
29 |
#include "classfile/systemDictionary.hpp"
|
|
30 |
#include "gc/shared/collectedHeap.hpp"
|
|
31 |
#include "interpreter/interpreter.hpp"
|
|
32 |
#include "oops/arrayOop.hpp"
|
|
33 |
#include "oops/markOop.hpp"
|
|
34 |
#include "runtime/basicLock.hpp"
|
|
35 |
#include "runtime/biasedLocking.hpp"
|
|
36 |
#include "runtime/os.hpp"
|
|
37 |
#include "runtime/stubRoutines.hpp"
|
|
38 |
#include "runtime/sharedRuntime.hpp"
|
|
39 |
|
|
40 |
|
|
41 |
void C1_MacroAssembler::inline_cache_check(Register receiver, Register iCache) {
|
|
42 |
const Register temp_reg = R12_scratch2;
|
|
43 |
verify_oop(receiver);
|
|
44 |
load_klass(temp_reg, receiver);
|
|
45 |
if (TrapBasedICMissChecks) {
|
|
46 |
trap_ic_miss_check(temp_reg, iCache);
|
|
47 |
} else {
|
|
48 |
Label L;
|
|
49 |
cmpd(CCR0, temp_reg, iCache);
|
|
50 |
beq(CCR0, L);
|
|
51 |
//load_const_optimized(temp_reg, SharedRuntime::get_ic_miss_stub(), R0);
|
|
52 |
calculate_address_from_global_toc(temp_reg, SharedRuntime::get_ic_miss_stub(), true, true, false);
|
|
53 |
mtctr(temp_reg);
|
|
54 |
bctr();
|
|
55 |
align(32, 12);
|
|
56 |
bind(L);
|
|
57 |
}
|
|
58 |
}
|
|
59 |
|
|
60 |
|
|
61 |
void C1_MacroAssembler::explicit_null_check(Register base) {
|
|
62 |
Unimplemented();
|
|
63 |
}
|
|
64 |
|
|
65 |
|
|
66 |
void C1_MacroAssembler::build_frame(int frame_size_in_bytes, int bang_size_in_bytes) {
|
|
67 |
assert(bang_size_in_bytes >= frame_size_in_bytes, "stack bang size incorrect");
|
|
68 |
// Make sure there is enough stack space for this method's activation.
|
|
69 |
generate_stack_overflow_check(bang_size_in_bytes);
|
|
70 |
|
|
71 |
// Create the frame.
|
|
72 |
const Register return_pc = R0;
|
|
73 |
|
|
74 |
mflr(return_pc);
|
|
75 |
// Get callers sp.
|
|
76 |
std(return_pc, _abi(lr), R1_SP); // SP->lr = return_pc
|
|
77 |
push_frame(frame_size_in_bytes, R0); // SP -= frame_size_in_bytes
|
|
78 |
}
|
|
79 |
|
|
80 |
|
|
81 |
void C1_MacroAssembler::unverified_entry(Register receiver, Register ic_klass) {
|
|
82 |
Unimplemented(); // Currently unused.
|
|
83 |
//if (C1Breakpoint) illtrap();
|
|
84 |
//inline_cache_check(receiver, ic_klass);
|
|
85 |
}
|
|
86 |
|
|
87 |
|
|
88 |
void C1_MacroAssembler::verified_entry() {
|
|
89 |
if (C1Breakpoint) illtrap();
|
|
90 |
// build frame
|
|
91 |
}
|
|
92 |
|
|
93 |
|
|
94 |
void C1_MacroAssembler::lock_object(Register Rmark, Register Roop, Register Rbox, Register Rscratch, Label& slow_case) {
|
|
95 |
assert_different_registers(Rmark, Roop, Rbox, Rscratch);
|
|
96 |
|
|
97 |
Label done, cas_failed, slow_int;
|
|
98 |
|
|
99 |
// The following move must be the first instruction of emitted since debug
|
|
100 |
// information may be generated for it.
|
|
101 |
// Load object header.
|
|
102 |
ld(Rmark, oopDesc::mark_offset_in_bytes(), Roop);
|
|
103 |
|
|
104 |
verify_oop(Roop);
|
|
105 |
|
|
106 |
// Save object being locked into the BasicObjectLock...
|
|
107 |
std(Roop, BasicObjectLock::obj_offset_in_bytes(), Rbox);
|
|
108 |
|
|
109 |
if (UseBiasedLocking) {
|
|
110 |
biased_locking_enter(CCR0, Roop, Rmark, Rscratch, R0, done, &slow_int);
|
|
111 |
}
|
|
112 |
|
|
113 |
// ... and mark it unlocked.
|
|
114 |
ori(Rmark, Rmark, markOopDesc::unlocked_value);
|
|
115 |
|
|
116 |
// Save unlocked object header into the displaced header location on the stack.
|
|
117 |
std(Rmark, BasicLock::displaced_header_offset_in_bytes(), Rbox);
|
|
118 |
|
|
119 |
// Compare object markOop with Rmark and if equal exchange Rscratch with object markOop.
|
|
120 |
assert(oopDesc::mark_offset_in_bytes() == 0, "cas must take a zero displacement");
|
|
121 |
cmpxchgd(/*flag=*/CCR0,
|
|
122 |
/*current_value=*/Rscratch,
|
|
123 |
/*compare_value=*/Rmark,
|
|
124 |
/*exchange_value=*/Rbox,
|
|
125 |
/*where=*/Roop/*+0==mark_offset_in_bytes*/,
|
|
126 |
MacroAssembler::MemBarRel | MacroAssembler::MemBarAcq,
|
|
127 |
MacroAssembler::cmpxchgx_hint_acquire_lock(),
|
|
128 |
noreg,
|
|
129 |
&cas_failed,
|
|
130 |
/*check without membar and ldarx first*/true);
|
|
131 |
// If compare/exchange succeeded we found an unlocked object and we now have locked it
|
|
132 |
// hence we are done.
|
|
133 |
b(done);
|
|
134 |
|
|
135 |
bind(slow_int);
|
|
136 |
b(slow_case); // far
|
|
137 |
|
|
138 |
bind(cas_failed);
|
|
139 |
// We did not find an unlocked object so see if this is a recursive case.
|
|
140 |
sub(Rscratch, Rscratch, R1_SP);
|
|
141 |
load_const_optimized(R0, (~(os::vm_page_size()-1) | markOopDesc::lock_mask_in_place));
|
|
142 |
and_(R0/*==0?*/, Rscratch, R0);
|
|
143 |
std(R0/*==0, perhaps*/, BasicLock::displaced_header_offset_in_bytes(), Rbox);
|
|
144 |
bne(CCR0, slow_int);
|
|
145 |
|
|
146 |
bind(done);
|
|
147 |
}
|
|
148 |
|
|
149 |
|
|
150 |
void C1_MacroAssembler::unlock_object(Register Rmark, Register Roop, Register Rbox, Label& slow_case) {
|
|
151 |
assert_different_registers(Rmark, Roop, Rbox);
|
|
152 |
|
|
153 |
Label slow_int, done;
|
|
154 |
|
|
155 |
Address mark_addr(Roop, oopDesc::mark_offset_in_bytes());
|
|
156 |
assert(mark_addr.disp() == 0, "cas must take a zero displacement");
|
|
157 |
|
|
158 |
if (UseBiasedLocking) {
|
|
159 |
// Load the object out of the BasicObjectLock.
|
|
160 |
ld(Roop, BasicObjectLock::obj_offset_in_bytes(), Rbox);
|
|
161 |
verify_oop(Roop);
|
|
162 |
biased_locking_exit(CCR0, Roop, R0, done);
|
|
163 |
}
|
|
164 |
// Test first it it is a fast recursive unlock.
|
|
165 |
ld(Rmark, BasicLock::displaced_header_offset_in_bytes(), Rbox);
|
|
166 |
cmpdi(CCR0, Rmark, 0);
|
|
167 |
beq(CCR0, done);
|
|
168 |
if (!UseBiasedLocking) {
|
|
169 |
// Load object.
|
|
170 |
ld(Roop, BasicObjectLock::obj_offset_in_bytes(), Rbox);
|
|
171 |
verify_oop(Roop);
|
|
172 |
}
|
|
173 |
|
|
174 |
// Check if it is still a light weight lock, this is is true if we see
|
|
175 |
// the stack address of the basicLock in the markOop of the object.
|
|
176 |
cmpxchgd(/*flag=*/CCR0,
|
|
177 |
/*current_value=*/R0,
|
|
178 |
/*compare_value=*/Rbox,
|
|
179 |
/*exchange_value=*/Rmark,
|
|
180 |
/*where=*/Roop,
|
|
181 |
MacroAssembler::MemBarRel,
|
|
182 |
MacroAssembler::cmpxchgx_hint_release_lock(),
|
|
183 |
noreg,
|
|
184 |
&slow_int);
|
|
185 |
b(done);
|
|
186 |
bind(slow_int);
|
|
187 |
b(slow_case); // far
|
|
188 |
|
|
189 |
// Done
|
|
190 |
bind(done);
|
|
191 |
}
|
|
192 |
|
|
193 |
|
|
194 |
void C1_MacroAssembler::try_allocate(
|
|
195 |
Register obj, // result: pointer to object after successful allocation
|
|
196 |
Register var_size_in_bytes, // object size in bytes if unknown at compile time; invalid otherwise
|
|
197 |
int con_size_in_bytes, // object size in bytes if known at compile time
|
|
198 |
Register t1, // temp register, must be global register for incr_allocated_bytes
|
|
199 |
Register t2, // temp register
|
|
200 |
Label& slow_case // continuation point if fast allocation fails
|
|
201 |
) {
|
|
202 |
if (UseTLAB) {
|
|
203 |
tlab_allocate(obj, var_size_in_bytes, con_size_in_bytes, t1, slow_case);
|
|
204 |
} else {
|
|
205 |
eden_allocate(obj, var_size_in_bytes, con_size_in_bytes, t1, t2, slow_case);
|
|
206 |
RegisterOrConstant size_in_bytes = var_size_in_bytes->is_valid()
|
|
207 |
? RegisterOrConstant(var_size_in_bytes)
|
|
208 |
: RegisterOrConstant(con_size_in_bytes);
|
|
209 |
incr_allocated_bytes(size_in_bytes, t1, t2);
|
|
210 |
}
|
|
211 |
}
|
|
212 |
|
|
213 |
|
|
214 |
void C1_MacroAssembler::initialize_header(Register obj, Register klass, Register len, Register t1, Register t2) {
|
|
215 |
assert_different_registers(obj, klass, len, t1, t2);
|
|
216 |
if (UseBiasedLocking && !len->is_valid()) {
|
|
217 |
ld(t1, in_bytes(Klass::prototype_header_offset()), klass);
|
|
218 |
} else {
|
|
219 |
load_const_optimized(t1, (intx)markOopDesc::prototype());
|
|
220 |
}
|
|
221 |
std(t1, oopDesc::mark_offset_in_bytes(), obj);
|
|
222 |
store_klass(obj, klass);
|
|
223 |
if (len->is_valid()) {
|
|
224 |
stw(len, arrayOopDesc::length_offset_in_bytes(), obj);
|
|
225 |
} else if (UseCompressedClassPointers) {
|
|
226 |
// Otherwise length is in the class gap.
|
|
227 |
store_klass_gap(obj);
|
|
228 |
}
|
|
229 |
}
|
|
230 |
|
|
231 |
|
|
232 |
void C1_MacroAssembler::initialize_body(Register base, Register index) {
|
|
233 |
assert_different_registers(base, index);
|
|
234 |
srdi(index, index, LogBytesPerWord);
|
|
235 |
clear_memory_doubleword(base, index);
|
|
236 |
}
|
|
237 |
|
|
238 |
void C1_MacroAssembler::initialize_body(Register obj, Register tmp1, Register tmp2,
|
|
239 |
int obj_size_in_bytes, int hdr_size_in_bytes) {
|
|
240 |
const int index = (obj_size_in_bytes - hdr_size_in_bytes) / HeapWordSize;
|
|
241 |
|
|
242 |
const int cl_size = VM_Version::L1_data_cache_line_size(),
|
|
243 |
cl_dwords = cl_size>>3,
|
|
244 |
cl_dw_addr_bits = exact_log2(cl_dwords);
|
|
245 |
|
|
246 |
const Register tmp = R0,
|
|
247 |
base_ptr = tmp1,
|
|
248 |
cnt_dwords = tmp2;
|
|
249 |
|
|
250 |
if (index <= 6) {
|
|
251 |
// Use explicit NULL stores.
|
|
252 |
if (index > 0) { li(tmp, 0); }
|
|
253 |
for (int i = 0; i < index; ++i) { std(tmp, hdr_size_in_bytes + i * HeapWordSize, obj); }
|
|
254 |
|
|
255 |
} else if (index < (2<<cl_dw_addr_bits)-1) {
|
|
256 |
// simple loop
|
|
257 |
Label loop;
|
|
258 |
|
|
259 |
li(cnt_dwords, index);
|
|
260 |
addi(base_ptr, obj, hdr_size_in_bytes); // Compute address of first element.
|
|
261 |
li(tmp, 0);
|
|
262 |
mtctr(cnt_dwords); // Load counter.
|
|
263 |
bind(loop);
|
|
264 |
std(tmp, 0, base_ptr); // Clear 8byte aligned block.
|
|
265 |
addi(base_ptr, base_ptr, 8);
|
|
266 |
bdnz(loop);
|
|
267 |
|
|
268 |
} else {
|
|
269 |
// like clear_memory_doubleword
|
|
270 |
Label startloop, fast, fastloop, restloop, done;
|
|
271 |
|
|
272 |
addi(base_ptr, obj, hdr_size_in_bytes); // Compute address of first element.
|
|
273 |
load_const_optimized(cnt_dwords, index);
|
|
274 |
rldicl_(tmp, base_ptr, 64-3, 64-cl_dw_addr_bits); // Extract dword offset within first cache line.
|
|
275 |
beq(CCR0, fast); // Already 128byte aligned.
|
|
276 |
|
|
277 |
subfic(tmp, tmp, cl_dwords);
|
|
278 |
mtctr(tmp); // Set ctr to hit 128byte boundary (0<ctr<cl_dwords).
|
|
279 |
subf(cnt_dwords, tmp, cnt_dwords); // rest.
|
|
280 |
li(tmp, 0);
|
|
281 |
|
|
282 |
bind(startloop); // Clear at the beginning to reach 128byte boundary.
|
|
283 |
std(tmp, 0, base_ptr); // Clear 8byte aligned block.
|
|
284 |
addi(base_ptr, base_ptr, 8);
|
|
285 |
bdnz(startloop);
|
|
286 |
|
|
287 |
bind(fast); // Clear 128byte blocks.
|
|
288 |
srdi(tmp, cnt_dwords, cl_dw_addr_bits); // Loop count for 128byte loop (>0).
|
|
289 |
andi(cnt_dwords, cnt_dwords, cl_dwords-1); // Rest in dwords.
|
|
290 |
mtctr(tmp); // Load counter.
|
|
291 |
|
|
292 |
bind(fastloop);
|
|
293 |
dcbz(base_ptr); // Clear 128byte aligned block.
|
|
294 |
addi(base_ptr, base_ptr, cl_size);
|
|
295 |
bdnz(fastloop);
|
|
296 |
|
|
297 |
cmpdi(CCR0, cnt_dwords, 0); // size 0?
|
|
298 |
beq(CCR0, done); // rest == 0
|
|
299 |
li(tmp, 0);
|
|
300 |
mtctr(cnt_dwords); // Load counter.
|
|
301 |
|
|
302 |
bind(restloop); // Clear rest.
|
|
303 |
std(tmp, 0, base_ptr); // Clear 8byte aligned block.
|
|
304 |
addi(base_ptr, base_ptr, 8);
|
|
305 |
bdnz(restloop);
|
|
306 |
|
|
307 |
bind(done);
|
|
308 |
}
|
|
309 |
}
|
|
310 |
|
|
311 |
void C1_MacroAssembler::allocate_object(
|
|
312 |
Register obj, // result: pointer to object after successful allocation
|
|
313 |
Register t1, // temp register
|
|
314 |
Register t2, // temp register
|
|
315 |
Register t3, // temp register
|
|
316 |
int hdr_size, // object header size in words
|
|
317 |
int obj_size, // object size in words
|
|
318 |
Register klass, // object klass
|
|
319 |
Label& slow_case // continuation point if fast allocation fails
|
|
320 |
) {
|
|
321 |
assert_different_registers(obj, t1, t2, t3, klass);
|
|
322 |
|
|
323 |
// allocate space & initialize header
|
|
324 |
if (!is_simm16(obj_size * wordSize)) {
|
|
325 |
// Would need to use extra register to load
|
|
326 |
// object size => go the slow case for now.
|
|
327 |
b(slow_case);
|
|
328 |
return;
|
|
329 |
}
|
|
330 |
try_allocate(obj, noreg, obj_size * wordSize, t2, t3, slow_case);
|
|
331 |
|
|
332 |
initialize_object(obj, klass, noreg, obj_size * HeapWordSize, t1, t2);
|
|
333 |
}
|
|
334 |
|
|
335 |
void C1_MacroAssembler::initialize_object(
|
|
336 |
Register obj, // result: pointer to object after successful allocation
|
|
337 |
Register klass, // object klass
|
|
338 |
Register var_size_in_bytes, // object size in bytes if unknown at compile time; invalid otherwise
|
|
339 |
int con_size_in_bytes, // object size in bytes if known at compile time
|
|
340 |
Register t1, // temp register
|
|
341 |
Register t2 // temp register
|
|
342 |
) {
|
|
343 |
const int hdr_size_in_bytes = instanceOopDesc::header_size() * HeapWordSize;
|
|
344 |
|
|
345 |
initialize_header(obj, klass, noreg, t1, t2);
|
|
346 |
|
|
347 |
#ifdef ASSERT
|
|
348 |
{
|
|
349 |
lwz(t1, in_bytes(Klass::layout_helper_offset()), klass);
|
|
350 |
if (var_size_in_bytes != noreg) {
|
|
351 |
cmpw(CCR0, t1, var_size_in_bytes);
|
|
352 |
} else {
|
|
353 |
cmpwi(CCR0, t1, con_size_in_bytes);
|
|
354 |
}
|
|
355 |
asm_assert_eq("bad size in initialize_object", 0x753);
|
|
356 |
}
|
|
357 |
#endif
|
|
358 |
|
|
359 |
// Initialize body.
|
|
360 |
if (var_size_in_bytes != noreg) {
|
|
361 |
// Use a loop.
|
|
362 |
addi(t1, obj, hdr_size_in_bytes); // Compute address of first element.
|
|
363 |
addi(t2, var_size_in_bytes, -hdr_size_in_bytes); // Compute size of body.
|
|
364 |
initialize_body(t1, t2);
|
|
365 |
} else if (con_size_in_bytes > hdr_size_in_bytes) {
|
|
366 |
// Use a loop.
|
|
367 |
initialize_body(obj, t1, t2, con_size_in_bytes, hdr_size_in_bytes);
|
|
368 |
}
|
|
369 |
|
|
370 |
if (CURRENT_ENV->dtrace_alloc_probes()) {
|
|
371 |
Unimplemented();
|
|
372 |
// assert(obj == O0, "must be");
|
|
373 |
// call(CAST_FROM_FN_PTR(address, Runtime1::entry_for(Runtime1::dtrace_object_alloc_id)),
|
|
374 |
// relocInfo::runtime_call_type);
|
|
375 |
}
|
|
376 |
|
|
377 |
verify_oop(obj);
|
|
378 |
}
|
|
379 |
|
|
380 |
|
|
381 |
void C1_MacroAssembler::allocate_array(
|
|
382 |
Register obj, // result: pointer to array after successful allocation
|
|
383 |
Register len, // array length
|
|
384 |
Register t1, // temp register
|
|
385 |
Register t2, // temp register
|
|
386 |
Register t3, // temp register
|
|
387 |
int hdr_size, // object header size in words
|
|
388 |
int elt_size, // element size in bytes
|
|
389 |
Register klass, // object klass
|
|
390 |
Label& slow_case // continuation point if fast allocation fails
|
|
391 |
) {
|
|
392 |
assert_different_registers(obj, len, t1, t2, t3, klass);
|
|
393 |
|
|
394 |
// Determine alignment mask.
|
|
395 |
assert(!(BytesPerWord & 1), "must be a multiple of 2 for masking code to work");
|
|
396 |
int log2_elt_size = exact_log2(elt_size);
|
|
397 |
|
|
398 |
// Check for negative or excessive length.
|
|
399 |
size_t max_length = max_array_allocation_length >> log2_elt_size;
|
|
400 |
if (UseTLAB) {
|
|
401 |
size_t max_tlab = align_size_up(ThreadLocalAllocBuffer::max_size() >> log2_elt_size, 64*K);
|
|
402 |
if (max_tlab < max_length) { max_length = max_tlab; }
|
|
403 |
}
|
|
404 |
load_const_optimized(t1, max_length);
|
|
405 |
cmpld(CCR0, len, t1);
|
|
406 |
bc_far_optimized(Assembler::bcondCRbiIs1, bi0(CCR0, Assembler::greater), slow_case);
|
|
407 |
|
|
408 |
// compute array size
|
|
409 |
// note: If 0 <= len <= max_length, len*elt_size + header + alignment is
|
|
410 |
// smaller or equal to the largest integer; also, since top is always
|
|
411 |
// aligned, we can do the alignment here instead of at the end address
|
|
412 |
// computation.
|
|
413 |
const Register arr_size = t1;
|
|
414 |
Register arr_len_in_bytes = len;
|
|
415 |
if (elt_size != 1) {
|
|
416 |
sldi(t1, len, log2_elt_size);
|
|
417 |
arr_len_in_bytes = t1;
|
|
418 |
}
|
|
419 |
addi(arr_size, arr_len_in_bytes, hdr_size * wordSize + MinObjAlignmentInBytesMask); // Add space for header & alignment.
|
|
420 |
clrrdi(arr_size, arr_size, LogMinObjAlignmentInBytes); // Align array size.
|
|
421 |
|
|
422 |
// Allocate space & initialize header.
|
|
423 |
if (UseTLAB) {
|
|
424 |
tlab_allocate(obj, arr_size, 0, t2, slow_case);
|
|
425 |
} else {
|
|
426 |
eden_allocate(obj, arr_size, 0, t2, t3, slow_case);
|
|
427 |
}
|
|
428 |
initialize_header(obj, klass, len, t2, t3);
|
|
429 |
|
|
430 |
// Initialize body.
|
|
431 |
const Register base = t2;
|
|
432 |
const Register index = t3;
|
|
433 |
addi(base, obj, hdr_size * wordSize); // compute address of first element
|
|
434 |
addi(index, arr_size, -(hdr_size * wordSize)); // compute index = number of bytes to clear
|
|
435 |
initialize_body(base, index);
|
|
436 |
|
|
437 |
if (CURRENT_ENV->dtrace_alloc_probes()) {
|
|
438 |
Unimplemented();
|
|
439 |
//assert(obj == O0, "must be");
|
|
440 |
//call(CAST_FROM_FN_PTR(address, Runtime1::entry_for(Runtime1::dtrace_object_alloc_id)),
|
|
441 |
// relocInfo::runtime_call_type);
|
|
442 |
}
|
|
443 |
|
|
444 |
verify_oop(obj);
|
|
445 |
}
|
|
446 |
|
|
447 |
|
|
448 |
#ifndef PRODUCT
|
|
449 |
|
|
450 |
void C1_MacroAssembler::verify_stack_oop(int stack_offset) {
|
|
451 |
verify_oop_addr((RegisterOrConstant)(stack_offset + STACK_BIAS), R1_SP, "broken oop in stack slot");
|
|
452 |
}
|
|
453 |
|
|
454 |
void C1_MacroAssembler::verify_not_null_oop(Register r) {
|
|
455 |
Label not_null;
|
|
456 |
cmpdi(CCR0, r, 0);
|
|
457 |
bne(CCR0, not_null);
|
|
458 |
stop("non-null oop required");
|
|
459 |
bind(not_null);
|
|
460 |
if (!VerifyOops) return;
|
|
461 |
verify_oop(r);
|
|
462 |
}
|
|
463 |
|
|
464 |
#endif // PRODUCT
|
|
465 |
|
|
466 |
void C1_MacroAssembler::null_check(Register r, Label* Lnull) {
|
|
467 |
if (TrapBasedNullChecks) { // SIGTRAP based
|
|
468 |
trap_null_check(r);
|
|
469 |
} else { // explicit
|
|
470 |
//const address exception_entry = Runtime1::entry_for(Runtime1::throw_null_pointer_exception_id);
|
|
471 |
assert(Lnull != NULL, "must have Label for explicit check");
|
|
472 |
cmpdi(CCR0, r, 0);
|
|
473 |
bc_far_optimized(Assembler::bcondCRbiIs1, bi0(CCR0, Assembler::equal), *Lnull);
|
|
474 |
}
|
|
475 |
}
|
|
476 |
|
|
477 |
address C1_MacroAssembler::call_c_with_frame_resize(address dest, int frame_resize) {
|
|
478 |
if (frame_resize) { resize_frame(-frame_resize, R0); }
|
|
479 |
#if defined(ABI_ELFv2)
|
|
480 |
address return_pc = call_c(dest, relocInfo::runtime_call_type);
|
|
481 |
#else
|
|
482 |
address return_pc = call_c(CAST_FROM_FN_PTR(FunctionDescriptor*, dest), relocInfo::runtime_call_type);
|
|
483 |
#endif
|
|
484 |
if (frame_resize) { resize_frame(frame_resize, R0); }
|
|
485 |
return return_pc;
|
|
486 |
}
|