42664
|
1 |
/*
|
|
2 |
* Copyright (c) 2008, 2016, Oracle and/or its affiliates. All rights reserved.
|
|
3 |
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
4 |
*
|
|
5 |
* This code is free software; you can redistribute it and/or modify it
|
|
6 |
* under the terms of the GNU General Public License version 2 only, as
|
|
7 |
* published by the Free Software Foundation.
|
|
8 |
*
|
|
9 |
* This code is distributed in the hope that it will be useful, but WITHOUT
|
|
10 |
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
11 |
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
12 |
* version 2 for more details (a copy is included in the LICENSE file that
|
|
13 |
* accompanied this code).
|
|
14 |
*
|
|
15 |
* You should have received a copy of the GNU General Public License version
|
|
16 |
* 2 along with this work; if not, write to the Free Software Foundation,
|
|
17 |
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
18 |
*
|
|
19 |
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
20 |
* or visit www.oracle.com if you need additional information or have any
|
|
21 |
* questions.
|
|
22 |
*
|
|
23 |
*/
|
|
24 |
|
|
25 |
#ifndef CPU_ARM_VM_NATIVEINST_ARM_64_HPP
|
|
26 |
#define CPU_ARM_VM_NATIVEINST_ARM_64_HPP
|
|
27 |
|
|
28 |
#include "asm/macroAssembler.hpp"
|
|
29 |
#include "code/codeCache.hpp"
|
|
30 |
#include "memory/allocation.hpp"
|
|
31 |
#include "runtime/icache.hpp"
|
|
32 |
#include "runtime/os.hpp"
|
|
33 |
|
|
34 |
// -------------------------------------------------------------------
|
|
35 |
|
|
36 |
// Some experimental projects extend the ARM back-end by implementing
|
|
37 |
// what the front-end usually assumes is a single native instruction
|
|
38 |
// with a sequence of instructions.
|
|
39 |
//
|
|
40 |
// The 'Raw' variants are the low level initial code (usually one
|
|
41 |
// instruction wide but some of them were already composed
|
|
42 |
// instructions). They should be used only by the back-end.
|
|
43 |
//
|
|
44 |
// The non-raw classes are the front-end entry point, hiding potential
|
|
45 |
// back-end extensions or the actual instructions size.
|
|
46 |
class NativeInstruction;
|
|
47 |
|
|
48 |
class RawNativeInstruction VALUE_OBJ_CLASS_SPEC {
|
|
49 |
public:
|
|
50 |
|
|
51 |
enum ARM_specific {
|
|
52 |
instruction_size = Assembler::InstructionSize,
|
|
53 |
instruction_size_in_bits = instruction_size * BitsPerByte,
|
|
54 |
};
|
|
55 |
|
|
56 |
// illegal instruction used by NativeJump::patch_verified_entry
|
|
57 |
static const int zombie_illegal_instruction = 0xd4000542; // hvc #42
|
|
58 |
|
|
59 |
address addr_at(int offset) const { return (address)this + offset; }
|
|
60 |
address instruction_address() const { return addr_at(0); }
|
|
61 |
address next_raw_instruction_address() const { return addr_at(instruction_size); }
|
|
62 |
|
|
63 |
static RawNativeInstruction* at(address address) {
|
|
64 |
return (RawNativeInstruction*)address;
|
|
65 |
}
|
|
66 |
|
|
67 |
RawNativeInstruction* next_raw() const {
|
|
68 |
return at(next_raw_instruction_address());
|
|
69 |
}
|
|
70 |
|
|
71 |
int encoding() const {
|
|
72 |
return *(int*)this;
|
|
73 |
}
|
|
74 |
|
|
75 |
void set_encoding(int value) {
|
|
76 |
int old = encoding();
|
|
77 |
if (old != value) {
|
|
78 |
*(int*)this = value;
|
|
79 |
ICache::invalidate_word((address)this);
|
|
80 |
}
|
|
81 |
}
|
|
82 |
|
|
83 |
bool is_nop() const { return encoding() == (int)0xd503201f; }
|
|
84 |
bool is_b() const { return (encoding() & 0xfc000000) == 0x14000000; } // unconditional branch
|
|
85 |
bool is_b_cond() const { return (encoding() & 0xff000010) == 0x54000000; } // conditional branch
|
|
86 |
bool is_bl() const { return (encoding() & 0xfc000000) == 0x94000000; }
|
|
87 |
bool is_br() const { return (encoding() & 0xfffffc1f) == 0xd61f0000; }
|
|
88 |
bool is_blr() const { return (encoding() & 0xfffffc1f) == 0xd63f0000; }
|
|
89 |
bool is_ldr_literal() const { return (encoding() & 0xff000000) == 0x58000000; }
|
|
90 |
bool is_adr_aligned() const { return (encoding() & 0xff000000) == 0x10000000; } // adr Xn, <label>, where label is aligned to 4 bytes (address of instruction).
|
|
91 |
bool is_adr_aligned_lr() const { return (encoding() & 0xff00001f) == 0x1000001e; } // adr LR, <label>, where label is aligned to 4 bytes (address of instruction).
|
|
92 |
|
|
93 |
bool is_ldr_str_gp_reg_unsigned_imm() const { return (encoding() & 0x3f000000) == 0x39000000; } // ldr/str{b, sb, h, sh, _w, sw} Rt, [Rn, #imm]
|
|
94 |
bool is_ldr_str_fp_reg_unsigned_imm() const { return (encoding() & 0x3f000000) == 0x3D000000; } // ldr/str Rt(SIMD), [Rn, #imm]
|
|
95 |
bool is_ldr_str_reg_unsigned_imm() const { return is_ldr_str_gp_reg_unsigned_imm() || is_ldr_str_fp_reg_unsigned_imm(); }
|
|
96 |
|
|
97 |
bool is_stp_preindex() const { return (encoding() & 0xffc00000) == 0xa9800000; } // stp Xt1, Xt2, [Xn, #imm]!
|
|
98 |
bool is_ldp_postindex() const { return (encoding() & 0xffc00000) == 0xa8c00000; } // ldp Xt1, Xt2, [Xn] #imm
|
|
99 |
bool is_mov_sp() const { return (encoding() & 0xfffffc00) == 0x91000000; } // mov <Xn|SP>, <Xm|SP>
|
|
100 |
bool is_movn() const { return (encoding() & 0x7f800000) == 0x12800000; }
|
|
101 |
bool is_movz() const { return (encoding() & 0x7f800000) == 0x52800000; }
|
|
102 |
bool is_movk() const { return (encoding() & 0x7f800000) == 0x72800000; }
|
|
103 |
bool is_orr_imm() const { return (encoding() & 0x7f800000) == 0x32000000; }
|
|
104 |
bool is_cmp_rr() const { return (encoding() & 0x7fe00000) == 0x6b000000; }
|
|
105 |
bool is_csel() const { return (encoding() & 0x7fe00000) == 0x1a800000; }
|
|
106 |
bool is_sub_shift() const { return (encoding() & 0x7f200000) == 0x4b000000; } // sub Rd, Rn, shift (Rm, imm)
|
|
107 |
bool is_mov() const { return (encoding() & 0x7fe0ffe0) == 0x2a0003e0; } // mov Rd, Rm (orr Rd, ZR, shift (Rm, 0))
|
|
108 |
bool is_tst() const { return (encoding() & 0x7f20001f) == 0x6a00001f; } // tst Rn, shift (Rm, imm) (ands ZR, Rn, shift(Rm, imm))
|
|
109 |
bool is_lsr_imm() const { return (encoding() & 0x7f807c00) == 0x53007c00; } // lsr Rd, Rn, imm (ubfm Rd, Rn, imm, 31/63)
|
|
110 |
|
|
111 |
bool is_far_jump() const { return is_ldr_literal() && next_raw()->is_br(); }
|
|
112 |
bool is_fat_call() const {
|
|
113 |
return
|
|
114 |
#ifdef COMPILER2
|
|
115 |
(is_blr() && next_raw()->is_b()) ||
|
|
116 |
#endif
|
|
117 |
(is_adr_aligned_lr() && next_raw()->is_br());
|
|
118 |
}
|
|
119 |
bool is_far_call() const {
|
|
120 |
return is_ldr_literal() && next_raw()->is_fat_call();
|
|
121 |
}
|
|
122 |
|
|
123 |
bool is_ic_near_call() const { return is_adr_aligned_lr() && next_raw()->is_b(); }
|
|
124 |
bool is_ic_far_call() const { return is_adr_aligned_lr() && next_raw()->is_ldr_literal() && next_raw()->next_raw()->is_br(); }
|
|
125 |
bool is_ic_call() const { return is_ic_near_call() || is_ic_far_call(); }
|
|
126 |
|
|
127 |
bool is_jump() const { return is_b() || is_far_jump(); }
|
|
128 |
bool is_call() const { return is_bl() || is_far_call() || is_ic_call(); }
|
|
129 |
bool is_branch() const { return is_b() || is_bl(); }
|
|
130 |
|
|
131 |
// c2 doesn't use fixed registers for safepoint poll address
|
|
132 |
bool is_safepoint_poll() const {
|
|
133 |
return true;
|
|
134 |
}
|
|
135 |
|
|
136 |
bool is_save_all_registers(const RawNativeInstruction** next) const {
|
|
137 |
const RawNativeInstruction* current = this;
|
|
138 |
|
|
139 |
if (!current->is_stp_preindex()) return false; current = current->next_raw();
|
|
140 |
for (int i = 28; i >= 0; i -= 2) {
|
|
141 |
if (!current->is_stp_preindex()) return false; current = current->next_raw();
|
|
142 |
}
|
|
143 |
|
|
144 |
if (!current->is_adr_aligned()) return false; current = current->next_raw();
|
|
145 |
if (!current->is_ldr_str_gp_reg_unsigned_imm()) return false; current = current->next_raw();
|
|
146 |
if (!current->is_ldr_str_gp_reg_unsigned_imm()) return false; current = current->next_raw();
|
|
147 |
|
|
148 |
*next = (RawNativeInstruction*) current;
|
|
149 |
return true;
|
|
150 |
}
|
|
151 |
|
|
152 |
bool is_restore_all_registers(const RawNativeInstruction** next) const {
|
|
153 |
const RawNativeInstruction* current = this;
|
|
154 |
|
|
155 |
for (int i = 0; i <= 28; i += 2) {
|
|
156 |
if (!current->is_ldp_postindex()) return false; current = current->next_raw();
|
|
157 |
}
|
|
158 |
if (!current->is_ldp_postindex()) return false; current = current->next_raw();
|
|
159 |
|
|
160 |
*next = (RawNativeInstruction*) current;
|
|
161 |
return true;
|
|
162 |
}
|
|
163 |
|
|
164 |
const RawNativeInstruction* skip_bind_literal() const {
|
|
165 |
const RawNativeInstruction* current = this;
|
|
166 |
if (((uintptr_t)current) % wordSize != 0) {
|
|
167 |
assert(current->is_nop(), "should be");
|
|
168 |
current = current->next_raw();
|
|
169 |
}
|
|
170 |
assert(((uintptr_t)current) % wordSize == 0, "should be"); // bound literal should be aligned
|
|
171 |
current = current->next_raw()->next_raw();
|
|
172 |
return current;
|
|
173 |
}
|
|
174 |
|
|
175 |
bool is_stop(const RawNativeInstruction** next) const {
|
|
176 |
const RawNativeInstruction* current = this;
|
|
177 |
|
|
178 |
if (!current->is_save_all_registers(¤t)) return false;
|
|
179 |
if (!current->is_ldr_literal()) return false; current = current->next_raw();
|
|
180 |
if (!current->is_mov_sp()) return false; current = current->next_raw();
|
|
181 |
if (!current->is_ldr_literal()) return false; current = current->next_raw();
|
|
182 |
if (!current->is_br()) return false; current = current->next_raw();
|
|
183 |
|
|
184 |
current = current->skip_bind_literal();
|
|
185 |
current = current->skip_bind_literal();
|
|
186 |
|
|
187 |
*next = (RawNativeInstruction*) current;
|
|
188 |
return true;
|
|
189 |
}
|
|
190 |
|
|
191 |
bool is_mov_slow(const RawNativeInstruction** next = NULL) const {
|
|
192 |
const RawNativeInstruction* current = this;
|
|
193 |
|
|
194 |
if (current->is_orr_imm()) {
|
|
195 |
current = current->next_raw();
|
|
196 |
|
|
197 |
} else if (current->is_movn() || current->is_movz()) {
|
|
198 |
current = current->next_raw();
|
|
199 |
int movkCount = 0;
|
|
200 |
while (current->is_movk()) {
|
|
201 |
movkCount++;
|
|
202 |
if (movkCount > 3) return false;
|
|
203 |
current = current->next_raw();
|
|
204 |
}
|
|
205 |
|
|
206 |
} else {
|
|
207 |
return false;
|
|
208 |
}
|
|
209 |
|
|
210 |
if (next != NULL) {
|
|
211 |
*next = (RawNativeInstruction*)current;
|
|
212 |
}
|
|
213 |
return true;
|
|
214 |
}
|
|
215 |
|
|
216 |
#ifdef ASSERT
|
|
217 |
void skip_verify_heapbase(const RawNativeInstruction** next) const {
|
|
218 |
const RawNativeInstruction* current = this;
|
|
219 |
|
|
220 |
if (CheckCompressedOops) {
|
|
221 |
if (!current->is_ldr_str_gp_reg_unsigned_imm()) return; current = current->next_raw();
|
|
222 |
if (!current->is_stp_preindex()) return; current = current->next_raw();
|
|
223 |
// NOTE: temporary workaround, remove with m6-01?
|
|
224 |
// skip saving condition flags
|
|
225 |
current = current->next_raw();
|
|
226 |
current = current->next_raw();
|
|
227 |
|
|
228 |
if (!current->is_mov_slow(¤t)) return;
|
|
229 |
if (!current->is_cmp_rr()) return; current = current->next_raw();
|
|
230 |
if (!current->is_b_cond()) return; current = current->next_raw();
|
|
231 |
if (!current->is_stop(¤t)) return;
|
|
232 |
|
|
233 |
#ifdef COMPILER2
|
|
234 |
if (current->is_nop()) current = current->next_raw();
|
|
235 |
#endif
|
|
236 |
// NOTE: temporary workaround, remove with m6-01?
|
|
237 |
// skip restoring condition flags
|
|
238 |
current = current->next_raw();
|
|
239 |
current = current->next_raw();
|
|
240 |
|
|
241 |
if (!current->is_ldp_postindex()) return; current = current->next_raw();
|
|
242 |
if (!current->is_ldr_str_gp_reg_unsigned_imm()) return; current = current->next_raw();
|
|
243 |
}
|
|
244 |
|
|
245 |
*next = (RawNativeInstruction*) current;
|
|
246 |
}
|
|
247 |
#endif // ASSERT
|
|
248 |
|
|
249 |
bool is_ldr_global_ptr(const RawNativeInstruction** next) const {
|
|
250 |
const RawNativeInstruction* current = this;
|
|
251 |
|
|
252 |
if (!current->is_mov_slow(¤t)) return false;
|
|
253 |
if (!current->is_ldr_str_gp_reg_unsigned_imm()) return false; current = current->next_raw();
|
|
254 |
|
|
255 |
*next = (RawNativeInstruction*) current;
|
|
256 |
return true;
|
|
257 |
}
|
|
258 |
|
|
259 |
void skip_verify_oop(const RawNativeInstruction** next) const {
|
|
260 |
const RawNativeInstruction* current = this;
|
|
261 |
|
|
262 |
if (VerifyOops) {
|
|
263 |
if (!current->is_save_all_registers(¤t)) return;
|
|
264 |
|
|
265 |
if (current->is_mov()) {
|
|
266 |
current = current->next_raw();
|
|
267 |
}
|
|
268 |
|
|
269 |
if (!current->is_mov_sp()) return; current = current->next_raw();
|
|
270 |
if (!current->is_ldr_literal()) return; current = current->next_raw();
|
|
271 |
if (!current->is_ldr_global_ptr(¤t)) return;
|
|
272 |
if (!current->is_blr()) return; current = current->next_raw();
|
|
273 |
if (!current->is_restore_all_registers(¤t)) return;
|
|
274 |
if (!current->is_b()) return; current = current->next_raw();
|
|
275 |
|
|
276 |
current = current->skip_bind_literal();
|
|
277 |
}
|
|
278 |
|
|
279 |
*next = (RawNativeInstruction*) current;
|
|
280 |
}
|
|
281 |
|
|
282 |
void skip_encode_heap_oop(const RawNativeInstruction** next) const {
|
|
283 |
const RawNativeInstruction* current = this;
|
|
284 |
|
|
285 |
assert (Universe::heap() != NULL, "java heap should be initialized");
|
|
286 |
#ifdef ASSERT
|
|
287 |
current->skip_verify_heapbase(¤t);
|
|
288 |
#endif // ASSERT
|
|
289 |
current->skip_verify_oop(¤t);
|
|
290 |
|
|
291 |
if (Universe::narrow_oop_base() == NULL) {
|
|
292 |
if (Universe::narrow_oop_shift() != 0) {
|
|
293 |
if (!current->is_lsr_imm()) return; current = current->next_raw();
|
|
294 |
} else {
|
|
295 |
if (current->is_mov()) {
|
|
296 |
current = current->next_raw();
|
|
297 |
}
|
|
298 |
}
|
|
299 |
} else {
|
|
300 |
if (!current->is_tst()) return; current = current->next_raw();
|
|
301 |
if (!current->is_csel()) return; current = current->next_raw();
|
|
302 |
if (!current->is_sub_shift()) return; current = current->next_raw();
|
|
303 |
if (Universe::narrow_oop_shift() != 0) {
|
|
304 |
if (!current->is_lsr_imm()) return; current = current->next_raw();
|
|
305 |
}
|
|
306 |
}
|
|
307 |
|
|
308 |
*next = (RawNativeInstruction*) current;
|
|
309 |
}
|
|
310 |
|
|
311 |
void verify();
|
|
312 |
|
|
313 |
// For unit tests
|
|
314 |
static void test() {}
|
|
315 |
|
|
316 |
private:
|
|
317 |
|
|
318 |
void check_bits_range(int bits, int scale, int low_bit) const {
|
|
319 |
assert((0 <= low_bit) && (0 < bits) && (low_bit + bits <= instruction_size_in_bits), "invalid bits range");
|
|
320 |
assert((0 <= scale) && (scale <= 4), "scale is out of range");
|
|
321 |
}
|
|
322 |
|
|
323 |
void set_imm(int imm_encoding, int bits, int low_bit) {
|
|
324 |
int imm_mask = right_n_bits(bits) << low_bit;
|
|
325 |
assert((imm_encoding & ~imm_mask) == 0, "invalid imm encoding");
|
|
326 |
set_encoding((encoding() & ~imm_mask) | imm_encoding);
|
|
327 |
}
|
|
328 |
|
|
329 |
protected:
|
|
330 |
|
|
331 |
// Returns signed immediate from [low_bit .. low_bit + bits - 1] bits of this instruction, scaled by given scale.
|
|
332 |
int get_signed_imm(int bits, int scale, int low_bit) const {
|
|
333 |
check_bits_range(bits, scale, low_bit);
|
|
334 |
int high_bits_to_clean = (instruction_size_in_bits - (low_bit + bits));
|
|
335 |
return encoding() << high_bits_to_clean >> (high_bits_to_clean + low_bit) << scale;
|
|
336 |
}
|
|
337 |
|
|
338 |
// Puts given signed immediate into the [low_bit .. low_bit + bits - 1] bits of this instruction.
|
|
339 |
void set_signed_imm(int value, int bits, int scale, int low_bit) {
|
|
340 |
set_imm(Assembler::encode_imm(value, bits, scale, low_bit), bits, low_bit);
|
|
341 |
}
|
|
342 |
|
|
343 |
// Returns unsigned immediate from [low_bit .. low_bit + bits - 1] bits of this instruction, scaled by given scale.
|
|
344 |
int get_unsigned_imm(int bits, int scale, int low_bit) const {
|
|
345 |
check_bits_range(bits, scale, low_bit);
|
|
346 |
return ((encoding() >> low_bit) & right_n_bits(bits)) << scale;
|
|
347 |
}
|
|
348 |
|
|
349 |
// Puts given unsigned immediate into the [low_bit .. low_bit + bits - 1] bits of this instruction.
|
|
350 |
void set_unsigned_imm(int value, int bits, int scale, int low_bit) {
|
|
351 |
set_imm(Assembler::encode_unsigned_imm(value, bits, scale, low_bit), bits, low_bit);
|
|
352 |
}
|
|
353 |
|
|
354 |
int get_signed_offset(int bits, int low_bit) const {
|
|
355 |
return get_signed_imm(bits, 2, low_bit);
|
|
356 |
}
|
|
357 |
|
|
358 |
void set_signed_offset(int offset, int bits, int low_bit) {
|
|
359 |
set_signed_imm(offset, bits, 2, low_bit);
|
|
360 |
}
|
|
361 |
};
|
|
362 |
|
|
363 |
inline RawNativeInstruction* rawNativeInstruction_at(address address) {
|
|
364 |
RawNativeInstruction* instr = RawNativeInstruction::at(address);
|
|
365 |
#ifdef ASSERT
|
|
366 |
instr->verify();
|
|
367 |
#endif // ASSERT
|
|
368 |
return instr;
|
|
369 |
}
|
|
370 |
|
|
371 |
// -------------------------------------------------------------------
|
|
372 |
|
|
373 |
// Load/store register (unsigned scaled immediate)
|
|
374 |
class NativeMovRegMem: public RawNativeInstruction {
|
|
375 |
private:
|
|
376 |
int get_offset_scale() const {
|
|
377 |
return get_unsigned_imm(2, 0, 30);
|
|
378 |
}
|
|
379 |
|
|
380 |
public:
|
|
381 |
int offset() const {
|
|
382 |
return get_unsigned_imm(12, get_offset_scale(), 10);
|
|
383 |
}
|
|
384 |
|
|
385 |
void set_offset(int x);
|
|
386 |
|
|
387 |
void add_offset_in_bytes(int add_offset) {
|
|
388 |
set_offset(offset() + add_offset);
|
|
389 |
}
|
|
390 |
};
|
|
391 |
|
|
392 |
inline NativeMovRegMem* nativeMovRegMem_at(address address) {
|
|
393 |
const RawNativeInstruction* instr = rawNativeInstruction_at(address);
|
|
394 |
|
|
395 |
#ifdef COMPILER1
|
|
396 |
// NOP required for C1 patching
|
|
397 |
if (instr->is_nop()) {
|
|
398 |
instr = instr->next_raw();
|
|
399 |
}
|
|
400 |
#endif
|
|
401 |
|
|
402 |
instr->skip_encode_heap_oop(&instr);
|
|
403 |
|
|
404 |
assert(instr->is_ldr_str_reg_unsigned_imm(), "must be");
|
|
405 |
return (NativeMovRegMem*)instr;
|
|
406 |
}
|
|
407 |
|
|
408 |
// -------------------------------------------------------------------
|
|
409 |
|
|
410 |
class NativeInstruction : public RawNativeInstruction {
|
|
411 |
public:
|
|
412 |
static NativeInstruction* at(address address) {
|
|
413 |
return (NativeInstruction*)address;
|
|
414 |
}
|
|
415 |
|
|
416 |
public:
|
|
417 |
// No need to consider indirections while parsing NativeInstruction
|
|
418 |
address next_instruction_address() const {
|
|
419 |
return next_raw_instruction_address();
|
|
420 |
}
|
|
421 |
|
|
422 |
// next() is no longer defined to avoid confusion.
|
|
423 |
//
|
|
424 |
// The front end and most classes except for those defined in nativeInst_arm
|
|
425 |
// or relocInfo_arm should only use next_instruction_address(), skipping
|
|
426 |
// over composed instruction and ignoring back-end extensions.
|
|
427 |
//
|
|
428 |
// The back-end can use next_raw() when it knows the instruction sequence
|
|
429 |
// and only wants to skip a single native instruction.
|
|
430 |
};
|
|
431 |
|
|
432 |
inline NativeInstruction* nativeInstruction_at(address address) {
|
|
433 |
NativeInstruction* instr = NativeInstruction::at(address);
|
|
434 |
#ifdef ASSERT
|
|
435 |
instr->verify();
|
|
436 |
#endif // ASSERT
|
|
437 |
return instr;
|
|
438 |
}
|
|
439 |
|
|
440 |
// -------------------------------------------------------------------
|
|
441 |
class NativeInstructionLdrLiteral: public NativeInstruction {
|
|
442 |
public:
|
|
443 |
address literal_address() {
|
|
444 |
address la = instruction_address() + get_signed_offset(19, 5);
|
|
445 |
assert(la != instruction_address(), "literal points to instruction");
|
|
446 |
return la;
|
|
447 |
}
|
|
448 |
|
|
449 |
address after_literal_address() {
|
|
450 |
return literal_address() + wordSize;
|
|
451 |
}
|
|
452 |
|
|
453 |
void set_literal_address(address addr, address pc) {
|
|
454 |
assert(is_ldr_literal(), "must be");
|
|
455 |
int opc = (encoding() >> 30) & 0x3;
|
|
456 |
assert (opc != 0b01 || addr == pc || ((uintx)addr & 7) == 0, "ldr target should be aligned");
|
|
457 |
set_signed_offset(addr - pc, 19, 5);
|
|
458 |
}
|
|
459 |
|
|
460 |
void set_literal_address(address addr) {
|
|
461 |
set_literal_address(addr, instruction_address());
|
|
462 |
}
|
|
463 |
|
|
464 |
address literal_value() {
|
|
465 |
return *(address*)literal_address();
|
|
466 |
}
|
|
467 |
|
|
468 |
void set_literal_value(address dest) {
|
|
469 |
*(address*)literal_address() = dest;
|
|
470 |
}
|
|
471 |
};
|
|
472 |
|
|
473 |
inline NativeInstructionLdrLiteral* nativeLdrLiteral_at(address address) {
|
|
474 |
assert(nativeInstruction_at(address)->is_ldr_literal(), "must be");
|
|
475 |
return (NativeInstructionLdrLiteral*)address;
|
|
476 |
}
|
|
477 |
|
|
478 |
// -------------------------------------------------------------------
|
|
479 |
// Common class for branch instructions with 26-bit immediate offset: B (unconditional) and BL
|
|
480 |
class NativeInstructionBranchImm26: public NativeInstruction {
|
|
481 |
public:
|
|
482 |
address destination(int adj = 0) const {
|
|
483 |
return instruction_address() + get_signed_offset(26, 0) + adj;
|
|
484 |
}
|
|
485 |
|
|
486 |
void set_destination(address dest) {
|
|
487 |
intptr_t offset = (intptr_t)(dest - instruction_address());
|
|
488 |
assert((offset & 0x3) == 0, "should be aligned");
|
|
489 |
set_signed_offset(offset, 26, 0);
|
|
490 |
}
|
|
491 |
};
|
|
492 |
|
|
493 |
inline NativeInstructionBranchImm26* nativeB_at(address address) {
|
|
494 |
assert(nativeInstruction_at(address)->is_b(), "must be");
|
|
495 |
return (NativeInstructionBranchImm26*)address;
|
|
496 |
}
|
|
497 |
|
|
498 |
inline NativeInstructionBranchImm26* nativeBL_at(address address) {
|
|
499 |
assert(nativeInstruction_at(address)->is_bl(), "must be");
|
|
500 |
return (NativeInstructionBranchImm26*)address;
|
|
501 |
}
|
|
502 |
|
|
503 |
// -------------------------------------------------------------------
|
|
504 |
class NativeInstructionAdrLR: public NativeInstruction {
|
|
505 |
public:
|
|
506 |
// Returns address which is loaded into LR by this instruction.
|
|
507 |
address target_lr_value() {
|
|
508 |
return instruction_address() + get_signed_offset(19, 5);
|
|
509 |
}
|
|
510 |
};
|
|
511 |
|
|
512 |
inline NativeInstructionAdrLR* nativeAdrLR_at(address address) {
|
|
513 |
assert(nativeInstruction_at(address)->is_adr_aligned_lr(), "must be");
|
|
514 |
return (NativeInstructionAdrLR*)address;
|
|
515 |
}
|
|
516 |
|
|
517 |
// -------------------------------------------------------------------
|
|
518 |
class RawNativeCall: public NativeInstruction {
|
|
519 |
public:
|
|
520 |
|
|
521 |
address return_address() const {
|
|
522 |
if (is_bl()) {
|
|
523 |
return next_raw_instruction_address();
|
|
524 |
|
|
525 |
} else if (is_far_call()) {
|
|
526 |
#ifdef COMPILER2
|
|
527 |
if (next_raw()->is_blr()) {
|
|
528 |
// ldr_literal; blr; ret_addr: b skip_literal;
|
|
529 |
return addr_at(2 * instruction_size);
|
|
530 |
}
|
|
531 |
#endif
|
|
532 |
assert(next_raw()->is_adr_aligned_lr() && next_raw()->next_raw()->is_br(), "must be");
|
|
533 |
return nativeLdrLiteral_at(instruction_address())->after_literal_address();
|
|
534 |
|
|
535 |
} else if (is_ic_call()) {
|
|
536 |
return nativeAdrLR_at(instruction_address())->target_lr_value();
|
|
537 |
|
|
538 |
} else {
|
|
539 |
ShouldNotReachHere();
|
|
540 |
return NULL;
|
|
541 |
}
|
|
542 |
}
|
|
543 |
|
|
544 |
address destination(int adj = 0) const {
|
|
545 |
if (is_bl()) {
|
|
546 |
return nativeBL_at(instruction_address())->destination(adj);
|
|
547 |
|
|
548 |
} else if (is_far_call()) {
|
|
549 |
return nativeLdrLiteral_at(instruction_address())->literal_value();
|
|
550 |
|
|
551 |
} else if (is_adr_aligned_lr()) {
|
|
552 |
RawNativeInstruction *next = next_raw();
|
|
553 |
if (next->is_b()) {
|
|
554 |
// ic_near_call
|
|
555 |
return nativeB_at(next->instruction_address())->destination(adj);
|
|
556 |
} else if (next->is_far_jump()) {
|
|
557 |
// ic_far_call
|
|
558 |
return nativeLdrLiteral_at(next->instruction_address())->literal_value();
|
|
559 |
}
|
|
560 |
}
|
|
561 |
ShouldNotReachHere();
|
|
562 |
return NULL;
|
|
563 |
}
|
|
564 |
|
|
565 |
void set_destination(address dest) {
|
|
566 |
if (is_bl()) {
|
|
567 |
nativeBL_at(instruction_address())->set_destination(dest);
|
|
568 |
return;
|
|
569 |
}
|
|
570 |
if (is_far_call()) {
|
|
571 |
nativeLdrLiteral_at(instruction_address())->set_literal_value(dest);
|
|
572 |
OrderAccess::storeload(); // overkill if caller holds lock?
|
|
573 |
return;
|
|
574 |
}
|
|
575 |
if (is_adr_aligned_lr()) {
|
|
576 |
RawNativeInstruction *next = next_raw();
|
|
577 |
if (next->is_b()) {
|
|
578 |
// ic_near_call
|
|
579 |
nativeB_at(next->instruction_address())->set_destination(dest);
|
|
580 |
return;
|
|
581 |
}
|
|
582 |
if (next->is_far_jump()) {
|
|
583 |
// ic_far_call
|
|
584 |
nativeLdrLiteral_at(next->instruction_address())->set_literal_value(dest);
|
|
585 |
OrderAccess::storeload(); // overkill if caller holds lock?
|
|
586 |
return;
|
|
587 |
}
|
|
588 |
}
|
|
589 |
ShouldNotReachHere();
|
|
590 |
}
|
|
591 |
|
|
592 |
void set_destination_mt_safe(address dest) {
|
|
593 |
assert(CodeCache::contains(dest), "call target should be from code cache (required by ic_call and patchable_call)");
|
|
594 |
set_destination(dest);
|
|
595 |
}
|
|
596 |
|
|
597 |
void verify() {
|
|
598 |
assert(RawNativeInstruction::is_call(), "should be");
|
|
599 |
}
|
|
600 |
|
|
601 |
void verify_alignment() {
|
|
602 |
// Nothing to do on ARM
|
|
603 |
}
|
|
604 |
};
|
|
605 |
|
|
606 |
inline RawNativeCall* rawNativeCall_at(address address) {
|
|
607 |
RawNativeCall * call = (RawNativeCall*)address;
|
|
608 |
call->verify();
|
|
609 |
return call;
|
|
610 |
}
|
|
611 |
|
|
612 |
class NativeCall: public RawNativeCall {
|
|
613 |
public:
|
|
614 |
|
|
615 |
// NativeCall::next_instruction_address() is used only to define the
|
|
616 |
// range where to look for the relocation information. We need not
|
|
617 |
// walk over composed instructions (as long as the relocation information
|
|
618 |
// is associated to the first instruction).
|
|
619 |
address next_instruction_address() const {
|
|
620 |
return next_raw_instruction_address();
|
|
621 |
}
|
|
622 |
|
|
623 |
static bool is_call_before(address return_address);
|
|
624 |
};
|
|
625 |
|
|
626 |
inline NativeCall* nativeCall_at(address address) {
|
|
627 |
NativeCall * call = (NativeCall*)address;
|
|
628 |
call->verify();
|
|
629 |
return call;
|
|
630 |
}
|
|
631 |
|
|
632 |
NativeCall* nativeCall_before(address return_address);
|
|
633 |
|
|
634 |
// -------------------------------------------------------------------
|
|
635 |
class NativeGeneralJump: public NativeInstruction {
|
|
636 |
public:
|
|
637 |
|
|
638 |
address jump_destination() const {
|
|
639 |
return nativeB_at(instruction_address())->destination();
|
|
640 |
}
|
|
641 |
|
|
642 |
static void replace_mt_safe(address instr_addr, address code_buffer);
|
|
643 |
|
|
644 |
static void insert_unconditional(address code_pos, address entry);
|
|
645 |
|
|
646 |
};
|
|
647 |
|
|
648 |
inline NativeGeneralJump* nativeGeneralJump_at(address address) {
|
|
649 |
assert(nativeInstruction_at(address)->is_b(), "must be");
|
|
650 |
return (NativeGeneralJump*)address;
|
|
651 |
}
|
|
652 |
|
|
653 |
// -------------------------------------------------------------------
|
|
654 |
class RawNativeJump: public NativeInstruction {
|
|
655 |
public:
|
|
656 |
|
|
657 |
address jump_destination(int adj = 0) const {
|
|
658 |
if (is_b()) {
|
|
659 |
address a = nativeB_at(instruction_address())->destination(adj);
|
|
660 |
// Jump destination -1 is encoded as a jump to self
|
|
661 |
if (a == instruction_address()) {
|
|
662 |
return (address)-1;
|
|
663 |
}
|
|
664 |
return a;
|
|
665 |
} else {
|
|
666 |
assert(is_far_jump(), "should be");
|
|
667 |
return nativeLdrLiteral_at(instruction_address())->literal_value();
|
|
668 |
}
|
|
669 |
}
|
|
670 |
|
|
671 |
void set_jump_destination(address dest) {
|
|
672 |
if (is_b()) {
|
|
673 |
// Jump destination -1 is encoded as a jump to self
|
|
674 |
if (dest == (address)-1) {
|
|
675 |
dest = instruction_address();
|
|
676 |
}
|
|
677 |
nativeB_at(instruction_address())->set_destination(dest);
|
|
678 |
} else {
|
|
679 |
assert(is_far_jump(), "should be");
|
|
680 |
nativeLdrLiteral_at(instruction_address())->set_literal_value(dest);
|
|
681 |
}
|
|
682 |
}
|
|
683 |
};
|
|
684 |
|
|
685 |
inline RawNativeJump* rawNativeJump_at(address address) {
|
|
686 |
assert(rawNativeInstruction_at(address)->is_jump(), "must be");
|
|
687 |
return (RawNativeJump*)address;
|
|
688 |
}
|
|
689 |
|
|
690 |
// -------------------------------------------------------------------
|
|
691 |
class NativeMovConstReg: public NativeInstruction {
|
|
692 |
|
|
693 |
NativeMovConstReg *adjust() const {
|
|
694 |
return (NativeMovConstReg *)adjust(this);
|
|
695 |
}
|
|
696 |
|
|
697 |
public:
|
|
698 |
|
|
699 |
static RawNativeInstruction *adjust(const RawNativeInstruction *ni) {
|
|
700 |
#ifdef COMPILER1
|
|
701 |
// NOP required for C1 patching
|
|
702 |
if (ni->is_nop()) {
|
|
703 |
return ni->next_raw();
|
|
704 |
}
|
|
705 |
#endif
|
|
706 |
return (RawNativeInstruction *)ni;
|
|
707 |
}
|
|
708 |
|
|
709 |
intptr_t _data() const;
|
|
710 |
void set_data(intptr_t x);
|
|
711 |
|
|
712 |
intptr_t data() const {
|
|
713 |
return adjust()->_data();
|
|
714 |
}
|
|
715 |
|
|
716 |
bool is_pc_relative() {
|
|
717 |
return adjust()->is_ldr_literal();
|
|
718 |
}
|
|
719 |
|
|
720 |
void _set_pc_relative_offset(address addr, address pc) {
|
|
721 |
assert(is_ldr_literal(), "must be");
|
|
722 |
nativeLdrLiteral_at(instruction_address())->set_literal_address(addr, pc);
|
|
723 |
}
|
|
724 |
|
|
725 |
void set_pc_relative_offset(address addr, address pc) {
|
|
726 |
NativeMovConstReg *ni = adjust();
|
|
727 |
int dest_adj = ni->instruction_address() - instruction_address();
|
|
728 |
ni->_set_pc_relative_offset(addr, pc + dest_adj);
|
|
729 |
}
|
|
730 |
|
|
731 |
address _next_instruction_address() const {
|
|
732 |
#ifdef COMPILER2
|
|
733 |
if (is_movz()) {
|
|
734 |
// narrow constant
|
|
735 |
RawNativeInstruction* ni = next_raw();
|
|
736 |
assert(ni->is_movk(), "movz;movk expected");
|
|
737 |
return ni->next_raw_instruction_address();
|
|
738 |
}
|
|
739 |
#endif
|
|
740 |
assert(is_ldr_literal(), "must be");
|
|
741 |
return NativeInstruction::next_raw_instruction_address();
|
|
742 |
}
|
|
743 |
|
|
744 |
address next_instruction_address() const {
|
|
745 |
return adjust()->_next_instruction_address();
|
|
746 |
}
|
|
747 |
};
|
|
748 |
|
|
749 |
inline NativeMovConstReg* nativeMovConstReg_at(address address) {
|
|
750 |
RawNativeInstruction* ni = rawNativeInstruction_at(address);
|
|
751 |
|
|
752 |
ni = NativeMovConstReg::adjust(ni);
|
|
753 |
|
|
754 |
assert(ni->is_mov_slow() || ni->is_ldr_literal(), "must be");
|
|
755 |
return (NativeMovConstReg*)address;
|
|
756 |
}
|
|
757 |
|
|
758 |
// -------------------------------------------------------------------
|
|
759 |
class NativeJump: public RawNativeJump {
|
|
760 |
public:
|
|
761 |
|
|
762 |
static void check_verified_entry_alignment(address entry, address verified_entry);
|
|
763 |
|
|
764 |
static void patch_verified_entry(address entry, address verified_entry, address dest);
|
|
765 |
};
|
|
766 |
|
|
767 |
inline NativeJump* nativeJump_at(address address) {
|
|
768 |
assert(nativeInstruction_at(address)->is_jump(), "must be");
|
|
769 |
return (NativeJump*)address;
|
|
770 |
}
|
|
771 |
|
|
772 |
#endif // CPU_ARM_VM_NATIVEINST_ARM_64_HPP
|