--- a/src/hotspot/cpu/aarch64/aarch64.ad Thu May 09 07:21:32 2019 -0400
+++ b/src/hotspot/cpu/aarch64/aarch64.ad Thu May 09 14:26:03 2019 +0200
@@ -1962,7 +1962,7 @@
st->print_cr("# MachUEPNode");
if (UseCompressedClassPointers) {
st->print_cr("\tldrw rscratch1, j_rarg0 + oopDesc::klass_offset_in_bytes()]\t# compressed klass");
- if (Universe::narrow_klass_shift() != 0) {
+ if (CompressedKlassPointers::shift() != 0) {
st->print_cr("\tdecode_klass_not_null rscratch1, rscratch1");
}
} else {
@@ -2183,7 +2183,7 @@
// Implicit_null_check optimization moves the Decode along with the
// memory operation back up before the NullCheck.
bool Matcher::narrow_oop_use_complex_address() {
- return Universe::narrow_oop_shift() == 0;
+ return CompressedOops::shift() == 0;
}
bool Matcher::narrow_klass_use_complex_address() {
@@ -2194,12 +2194,12 @@
bool Matcher::const_oop_prefer_decode() {
// Prefer ConN+DecodeN over ConP in simple compressed oops mode.
- return Universe::narrow_oop_base() == NULL;
+ return CompressedOops::base() == NULL;
}
bool Matcher::const_klass_prefer_decode() {
// Prefer ConNKlass+DecodeNKlass over ConP in simple compressed klass mode.
- return Universe::narrow_klass_base() == NULL;
+ return CompressedKlassPointers::base() == NULL;
}
// Is it better to copy float constants, or load them directly from
@@ -5044,7 +5044,7 @@
operand indirectN(iRegN reg)
%{
- predicate(Universe::narrow_oop_shift() == 0);
+ predicate(CompressedOops::shift() == 0);
constraint(ALLOC_IN_RC(ptr_reg));
match(DecodeN reg);
op_cost(0);
@@ -5059,7 +5059,7 @@
operand indIndexScaledI2LN(iRegN reg, iRegI ireg, immIScale scale)
%{
- predicate(Universe::narrow_oop_shift() == 0 && size_fits_all_mem_uses(n->as_AddP(), n->in(AddPNode::Offset)->in(2)->get_int()));
+ predicate(CompressedOops::shift() == 0 && size_fits_all_mem_uses(n->as_AddP(), n->in(AddPNode::Offset)->in(2)->get_int()));
constraint(ALLOC_IN_RC(ptr_reg));
match(AddP (DecodeN reg) (LShiftL (ConvI2L ireg) scale));
op_cost(0);
@@ -5074,7 +5074,7 @@
operand indIndexScaledN(iRegN reg, iRegL lreg, immIScale scale)
%{
- predicate(Universe::narrow_oop_shift() == 0 && size_fits_all_mem_uses(n->as_AddP(), n->in(AddPNode::Offset)->in(2)->get_int()));
+ predicate(CompressedOops::shift() == 0 && size_fits_all_mem_uses(n->as_AddP(), n->in(AddPNode::Offset)->in(2)->get_int()));
constraint(ALLOC_IN_RC(ptr_reg));
match(AddP (DecodeN reg) (LShiftL lreg scale));
op_cost(0);
@@ -5089,7 +5089,7 @@
operand indIndexI2LN(iRegN reg, iRegI ireg)
%{
- predicate(Universe::narrow_oop_shift() == 0);
+ predicate(CompressedOops::shift() == 0);
constraint(ALLOC_IN_RC(ptr_reg));
match(AddP (DecodeN reg) (ConvI2L ireg));
op_cost(0);
@@ -5104,7 +5104,7 @@
operand indIndexN(iRegN reg, iRegL lreg)
%{
- predicate(Universe::narrow_oop_shift() == 0);
+ predicate(CompressedOops::shift() == 0);
constraint(ALLOC_IN_RC(ptr_reg));
match(AddP (DecodeN reg) lreg);
op_cost(0);
@@ -5119,7 +5119,7 @@
operand indOffIN(iRegN reg, immIOffset off)
%{
- predicate(Universe::narrow_oop_shift() == 0);
+ predicate(CompressedOops::shift() == 0);
constraint(ALLOC_IN_RC(ptr_reg));
match(AddP (DecodeN reg) off);
op_cost(0);
@@ -5134,7 +5134,7 @@
operand indOffLN(iRegN reg, immLoffset off)
%{
- predicate(Universe::narrow_oop_shift() == 0);
+ predicate(CompressedOops::shift() == 0);
constraint(ALLOC_IN_RC(ptr_reg));
match(AddP (DecodeN reg) off);
op_cost(0);
@@ -7039,8 +7039,8 @@
instruct storeImmN0(iRegIHeapbase heapbase, immN0 zero, memory mem)
%{
match(Set mem (StoreN mem zero));
- predicate(Universe::narrow_oop_base() == NULL &&
- Universe::narrow_klass_base() == NULL &&
+ predicate(CompressedOops::base() == NULL &&
+ CompressedKlassPointers::base() == NULL &&
(!needs_releasing_store(n)));
ins_cost(INSN_COST);
@@ -7822,7 +7822,7 @@
// in case of 32bit oops (heap < 4Gb).
instruct convN2I(iRegINoSp dst, iRegN src)
%{
- predicate(Universe::narrow_oop_shift() == 0);
+ predicate(CompressedOops::shift() == 0);
match(Set dst (ConvL2I (CastP2X (DecodeN src))));
ins_cost(INSN_COST);