--- /dev/null Thu Jan 01 00:00:00 1970 +0000
+++ b/src/hotspot/share/oops/methodData.cpp Tue Sep 12 19:03:39 2017 +0200
@@ -0,0 +1,1801 @@
+/*
+ * Copyright (c) 2000, 2017, Oracle and/or its affiliates. All rights reserved.
+ * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
+ *
+ * This code is free software; you can redistribute it and/or modify it
+ * under the terms of the GNU General Public License version 2 only, as
+ * published by the Free Software Foundation.
+ *
+ * This code is distributed in the hope that it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
+ * version 2 for more details (a copy is included in the LICENSE file that
+ * accompanied this code).
+ *
+ * You should have received a copy of the GNU General Public License version
+ * 2 along with this work; if not, write to the Free Software Foundation,
+ * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
+ * or visit www.oracle.com if you need additional information or have any
+ * questions.
+ *
+ */
+
+#include "precompiled.hpp"
+#include "classfile/systemDictionary.hpp"
+#include "compiler/compilerOracle.hpp"
+#include "interpreter/bytecode.hpp"
+#include "interpreter/bytecodeStream.hpp"
+#include "interpreter/linkResolver.hpp"
+#include "memory/heapInspection.hpp"
+#include "memory/metaspaceClosure.hpp"
+#include "memory/resourceArea.hpp"
+#include "oops/methodData.hpp"
+#include "prims/jvmtiRedefineClasses.hpp"
+#include "runtime/arguments.hpp"
+#include "runtime/compilationPolicy.hpp"
+#include "runtime/deoptimization.hpp"
+#include "runtime/handles.inline.hpp"
+#include "runtime/orderAccess.inline.hpp"
+#include "utilities/align.hpp"
+#include "utilities/copy.hpp"
+
+// ==================================================================
+// DataLayout
+//
+// Overlay for generic profiling data.
+
+// Some types of data layouts need a length field.
+bool DataLayout::needs_array_len(u1 tag) {
+ return (tag == multi_branch_data_tag) || (tag == arg_info_data_tag) || (tag == parameters_type_data_tag);
+}
+
+// Perform generic initialization of the data. More specific
+// initialization occurs in overrides of ProfileData::post_initialize.
+void DataLayout::initialize(u1 tag, u2 bci, int cell_count) {
+ _header._bits = (intptr_t)0;
+ _header._struct._tag = tag;
+ _header._struct._bci = bci;
+ for (int i = 0; i < cell_count; i++) {
+ set_cell_at(i, (intptr_t)0);
+ }
+ if (needs_array_len(tag)) {
+ set_cell_at(ArrayData::array_len_off_set, cell_count - 1); // -1 for header.
+ }
+ if (tag == call_type_data_tag) {
+ CallTypeData::initialize(this, cell_count);
+ } else if (tag == virtual_call_type_data_tag) {
+ VirtualCallTypeData::initialize(this, cell_count);
+ }
+}
+
+void DataLayout::clean_weak_klass_links(BoolObjectClosure* cl) {
+ ResourceMark m;
+ data_in()->clean_weak_klass_links(cl);
+}
+
+
+// ==================================================================
+// ProfileData
+//
+// A ProfileData object is created to refer to a section of profiling
+// data in a structured way.
+
+// Constructor for invalid ProfileData.
+ProfileData::ProfileData() {
+ _data = NULL;
+}
+
+char* ProfileData::print_data_on_helper(const MethodData* md) const {
+ DataLayout* dp = md->extra_data_base();
+ DataLayout* end = md->args_data_limit();
+ stringStream ss;
+ for (;; dp = MethodData::next_extra(dp)) {
+ assert(dp < end, "moved past end of extra data");
+ switch(dp->tag()) {
+ case DataLayout::speculative_trap_data_tag:
+ if (dp->bci() == bci()) {
+ SpeculativeTrapData* data = new SpeculativeTrapData(dp);
+ int trap = data->trap_state();
+ char buf[100];
+ ss.print("trap/");
+ data->method()->print_short_name(&ss);
+ ss.print("(%s) ", Deoptimization::format_trap_state(buf, sizeof(buf), trap));
+ }
+ break;
+ case DataLayout::bit_data_tag:
+ break;
+ case DataLayout::no_tag:
+ case DataLayout::arg_info_data_tag:
+ return ss.as_string();
+ break;
+ default:
+ fatal("unexpected tag %d", dp->tag());
+ }
+ }
+ return NULL;
+}
+
+void ProfileData::print_data_on(outputStream* st, const MethodData* md) const {
+ print_data_on(st, print_data_on_helper(md));
+}
+
+void ProfileData::print_shared(outputStream* st, const char* name, const char* extra) const {
+ st->print("bci: %d", bci());
+ st->fill_to(tab_width_one);
+ st->print("%s", name);
+ tab(st);
+ int trap = trap_state();
+ if (trap != 0) {
+ char buf[100];
+ st->print("trap(%s) ", Deoptimization::format_trap_state(buf, sizeof(buf), trap));
+ }
+ if (extra != NULL) {
+ st->print("%s", extra);
+ }
+ int flags = data()->flags();
+ if (flags != 0) {
+ st->print("flags(%d) ", flags);
+ }
+}
+
+void ProfileData::tab(outputStream* st, bool first) const {
+ st->fill_to(first ? tab_width_one : tab_width_two);
+}
+
+// ==================================================================
+// BitData
+//
+// A BitData corresponds to a one-bit flag. This is used to indicate
+// whether a checkcast bytecode has seen a null value.
+
+
+void BitData::print_data_on(outputStream* st, const char* extra) const {
+ print_shared(st, "BitData", extra);
+ st->cr();
+}
+
+// ==================================================================
+// CounterData
+//
+// A CounterData corresponds to a simple counter.
+
+void CounterData::print_data_on(outputStream* st, const char* extra) const {
+ print_shared(st, "CounterData", extra);
+ st->print_cr("count(%u)", count());
+}
+
+// ==================================================================
+// JumpData
+//
+// A JumpData is used to access profiling information for a direct
+// branch. It is a counter, used for counting the number of branches,
+// plus a data displacement, used for realigning the data pointer to
+// the corresponding target bci.
+
+void JumpData::post_initialize(BytecodeStream* stream, MethodData* mdo) {
+ assert(stream->bci() == bci(), "wrong pos");
+ int target;
+ Bytecodes::Code c = stream->code();
+ if (c == Bytecodes::_goto_w || c == Bytecodes::_jsr_w) {
+ target = stream->dest_w();
+ } else {
+ target = stream->dest();
+ }
+ int my_di = mdo->dp_to_di(dp());
+ int target_di = mdo->bci_to_di(target);
+ int offset = target_di - my_di;
+ set_displacement(offset);
+}
+
+void JumpData::print_data_on(outputStream* st, const char* extra) const {
+ print_shared(st, "JumpData", extra);
+ st->print_cr("taken(%u) displacement(%d)", taken(), displacement());
+}
+
+int TypeStackSlotEntries::compute_cell_count(Symbol* signature, bool include_receiver, int max) {
+ // Parameter profiling include the receiver
+ int args_count = include_receiver ? 1 : 0;
+ ResourceMark rm;
+ SignatureStream ss(signature);
+ args_count += ss.reference_parameter_count();
+ args_count = MIN2(args_count, max);
+ return args_count * per_arg_cell_count;
+}
+
+int TypeEntriesAtCall::compute_cell_count(BytecodeStream* stream) {
+ assert(Bytecodes::is_invoke(stream->code()), "should be invoke");
+ assert(TypeStackSlotEntries::per_arg_count() > ReturnTypeEntry::static_cell_count(), "code to test for arguments/results broken");
+ const methodHandle m = stream->method();
+ int bci = stream->bci();
+ Bytecode_invoke inv(m, bci);
+ int args_cell = 0;
+ if (MethodData::profile_arguments_for_invoke(m, bci)) {
+ args_cell = TypeStackSlotEntries::compute_cell_count(inv.signature(), false, TypeProfileArgsLimit);
+ }
+ int ret_cell = 0;
+ if (MethodData::profile_return_for_invoke(m, bci) && (inv.result_type() == T_OBJECT || inv.result_type() == T_ARRAY)) {
+ ret_cell = ReturnTypeEntry::static_cell_count();
+ }
+ int header_cell = 0;
+ if (args_cell + ret_cell > 0) {
+ header_cell = header_cell_count();
+ }
+
+ return header_cell + args_cell + ret_cell;
+}
+
+class ArgumentOffsetComputer : public SignatureInfo {
+private:
+ int _max;
+ GrowableArray<int> _offsets;
+
+ void set(int size, BasicType type) { _size += size; }
+ void do_object(int begin, int end) {
+ if (_offsets.length() < _max) {
+ _offsets.push(_size);
+ }
+ SignatureInfo::do_object(begin, end);
+ }
+ void do_array (int begin, int end) {
+ if (_offsets.length() < _max) {
+ _offsets.push(_size);
+ }
+ SignatureInfo::do_array(begin, end);
+ }
+
+public:
+ ArgumentOffsetComputer(Symbol* signature, int max)
+ : SignatureInfo(signature), _max(max), _offsets(Thread::current(), max) {
+ }
+
+ int total() { lazy_iterate_parameters(); return _size; }
+
+ int off_at(int i) const { return _offsets.at(i); }
+};
+
+void TypeStackSlotEntries::post_initialize(Symbol* signature, bool has_receiver, bool include_receiver) {
+ ResourceMark rm;
+ int start = 0;
+ // Parameter profiling include the receiver
+ if (include_receiver && has_receiver) {
+ set_stack_slot(0, 0);
+ set_type(0, type_none());
+ start += 1;
+ }
+ ArgumentOffsetComputer aos(signature, _number_of_entries-start);
+ aos.total();
+ for (int i = start; i < _number_of_entries; i++) {
+ set_stack_slot(i, aos.off_at(i-start) + (has_receiver ? 1 : 0));
+ set_type(i, type_none());
+ }
+}
+
+void CallTypeData::post_initialize(BytecodeStream* stream, MethodData* mdo) {
+ assert(Bytecodes::is_invoke(stream->code()), "should be invoke");
+ Bytecode_invoke inv(stream->method(), stream->bci());
+
+ SignatureStream ss(inv.signature());
+ if (has_arguments()) {
+#ifdef ASSERT
+ ResourceMark rm;
+ int count = MIN2(ss.reference_parameter_count(), (int)TypeProfileArgsLimit);
+ assert(count > 0, "room for args type but none found?");
+ check_number_of_arguments(count);
+#endif
+ _args.post_initialize(inv.signature(), inv.has_receiver(), false);
+ }
+
+ if (has_return()) {
+ assert(inv.result_type() == T_OBJECT || inv.result_type() == T_ARRAY, "room for a ret type but doesn't return obj?");
+ _ret.post_initialize();
+ }
+}
+
+void VirtualCallTypeData::post_initialize(BytecodeStream* stream, MethodData* mdo) {
+ assert(Bytecodes::is_invoke(stream->code()), "should be invoke");
+ Bytecode_invoke inv(stream->method(), stream->bci());
+
+ if (has_arguments()) {
+#ifdef ASSERT
+ ResourceMark rm;
+ SignatureStream ss(inv.signature());
+ int count = MIN2(ss.reference_parameter_count(), (int)TypeProfileArgsLimit);
+ assert(count > 0, "room for args type but none found?");
+ check_number_of_arguments(count);
+#endif
+ _args.post_initialize(inv.signature(), inv.has_receiver(), false);
+ }
+
+ if (has_return()) {
+ assert(inv.result_type() == T_OBJECT || inv.result_type() == T_ARRAY, "room for a ret type but doesn't return obj?");
+ _ret.post_initialize();
+ }
+}
+
+bool TypeEntries::is_loader_alive(BoolObjectClosure* is_alive_cl, intptr_t p) {
+ Klass* k = (Klass*)klass_part(p);
+ return k != NULL && k->is_loader_alive(is_alive_cl);
+}
+
+void TypeStackSlotEntries::clean_weak_klass_links(BoolObjectClosure* is_alive_cl) {
+ for (int i = 0; i < _number_of_entries; i++) {
+ intptr_t p = type(i);
+ if (!is_loader_alive(is_alive_cl, p)) {
+ set_type(i, with_status((Klass*)NULL, p));
+ }
+ }
+}
+
+void ReturnTypeEntry::clean_weak_klass_links(BoolObjectClosure* is_alive_cl) {
+ intptr_t p = type();
+ if (!is_loader_alive(is_alive_cl, p)) {
+ set_type(with_status((Klass*)NULL, p));
+ }
+}
+
+bool TypeEntriesAtCall::return_profiling_enabled() {
+ return MethodData::profile_return();
+}
+
+bool TypeEntriesAtCall::arguments_profiling_enabled() {
+ return MethodData::profile_arguments();
+}
+
+void TypeEntries::print_klass(outputStream* st, intptr_t k) {
+ if (is_type_none(k)) {
+ st->print("none");
+ } else if (is_type_unknown(k)) {
+ st->print("unknown");
+ } else {
+ valid_klass(k)->print_value_on(st);
+ }
+ if (was_null_seen(k)) {
+ st->print(" (null seen)");
+ }
+}
+
+void TypeStackSlotEntries::print_data_on(outputStream* st) const {
+ for (int i = 0; i < _number_of_entries; i++) {
+ _pd->tab(st);
+ st->print("%d: stack(%u) ", i, stack_slot(i));
+ print_klass(st, type(i));
+ st->cr();
+ }
+}
+
+void ReturnTypeEntry::print_data_on(outputStream* st) const {
+ _pd->tab(st);
+ print_klass(st, type());
+ st->cr();
+}
+
+void CallTypeData::print_data_on(outputStream* st, const char* extra) const {
+ CounterData::print_data_on(st, extra);
+ if (has_arguments()) {
+ tab(st, true);
+ st->print("argument types");
+ _args.print_data_on(st);
+ }
+ if (has_return()) {
+ tab(st, true);
+ st->print("return type");
+ _ret.print_data_on(st);
+ }
+}
+
+void VirtualCallTypeData::print_data_on(outputStream* st, const char* extra) const {
+ VirtualCallData::print_data_on(st, extra);
+ if (has_arguments()) {
+ tab(st, true);
+ st->print("argument types");
+ _args.print_data_on(st);
+ }
+ if (has_return()) {
+ tab(st, true);
+ st->print("return type");
+ _ret.print_data_on(st);
+ }
+}
+
+// ==================================================================
+// ReceiverTypeData
+//
+// A ReceiverTypeData is used to access profiling information about a
+// dynamic type check. It consists of a counter which counts the total times
+// that the check is reached, and a series of (Klass*, count) pairs
+// which are used to store a type profile for the receiver of the check.
+
+void ReceiverTypeData::clean_weak_klass_links(BoolObjectClosure* is_alive_cl) {
+ for (uint row = 0; row < row_limit(); row++) {
+ Klass* p = receiver(row);
+ if (p != NULL && !p->is_loader_alive(is_alive_cl)) {
+ clear_row(row);
+ }
+ }
+}
+
+#if INCLUDE_JVMCI
+void VirtualCallData::clean_weak_klass_links(BoolObjectClosure* is_alive_cl) {
+ ReceiverTypeData::clean_weak_klass_links(is_alive_cl);
+ for (uint row = 0; row < method_row_limit(); row++) {
+ Method* p = method(row);
+ if (p != NULL && !p->method_holder()->is_loader_alive(is_alive_cl)) {
+ clear_method_row(row);
+ }
+ }
+}
+
+void VirtualCallData::clean_weak_method_links() {
+ ReceiverTypeData::clean_weak_method_links();
+ for (uint row = 0; row < method_row_limit(); row++) {
+ Method* p = method(row);
+ if (p != NULL && !p->on_stack()) {
+ clear_method_row(row);
+ }
+ }
+}
+#endif // INCLUDE_JVMCI
+
+void ReceiverTypeData::print_receiver_data_on(outputStream* st) const {
+ uint row;
+ int entries = 0;
+ for (row = 0; row < row_limit(); row++) {
+ if (receiver(row) != NULL) entries++;
+ }
+#if INCLUDE_JVMCI
+ st->print_cr("count(%u) nonprofiled_count(%u) entries(%u)", count(), nonprofiled_count(), entries);
+#else
+ st->print_cr("count(%u) entries(%u)", count(), entries);
+#endif
+ int total = count();
+ for (row = 0; row < row_limit(); row++) {
+ if (receiver(row) != NULL) {
+ total += receiver_count(row);
+ }
+ }
+ for (row = 0; row < row_limit(); row++) {
+ if (receiver(row) != NULL) {
+ tab(st);
+ receiver(row)->print_value_on(st);
+ st->print_cr("(%u %4.2f)", receiver_count(row), (float) receiver_count(row) / (float) total);
+ }
+ }
+}
+void ReceiverTypeData::print_data_on(outputStream* st, const char* extra) const {
+ print_shared(st, "ReceiverTypeData", extra);
+ print_receiver_data_on(st);
+}
+
+#if INCLUDE_JVMCI
+void VirtualCallData::print_method_data_on(outputStream* st) const {
+ uint row;
+ int entries = 0;
+ for (row = 0; row < method_row_limit(); row++) {
+ if (method(row) != NULL) entries++;
+ }
+ tab(st);
+ st->print_cr("method_entries(%u)", entries);
+ int total = count();
+ for (row = 0; row < method_row_limit(); row++) {
+ if (method(row) != NULL) {
+ total += method_count(row);
+ }
+ }
+ for (row = 0; row < method_row_limit(); row++) {
+ if (method(row) != NULL) {
+ tab(st);
+ method(row)->print_value_on(st);
+ st->print_cr("(%u %4.2f)", method_count(row), (float) method_count(row) / (float) total);
+ }
+ }
+}
+#endif // INCLUDE_JVMCI
+
+void VirtualCallData::print_data_on(outputStream* st, const char* extra) const {
+ print_shared(st, "VirtualCallData", extra);
+ print_receiver_data_on(st);
+ print_method_data_on(st);
+}
+
+// ==================================================================
+// RetData
+//
+// A RetData is used to access profiling information for a ret bytecode.
+// It is composed of a count of the number of times that the ret has
+// been executed, followed by a series of triples of the form
+// (bci, count, di) which count the number of times that some bci was the
+// target of the ret and cache a corresponding displacement.
+
+void RetData::post_initialize(BytecodeStream* stream, MethodData* mdo) {
+ for (uint row = 0; row < row_limit(); row++) {
+ set_bci_displacement(row, -1);
+ set_bci(row, no_bci);
+ }
+ // release so other threads see a consistent state. bci is used as
+ // a valid flag for bci_displacement.
+ OrderAccess::release();
+}
+
+// This routine needs to atomically update the RetData structure, so the
+// caller needs to hold the RetData_lock before it gets here. Since taking
+// the lock can block (and allow GC) and since RetData is a ProfileData is a
+// wrapper around a derived oop, taking the lock in _this_ method will
+// basically cause the 'this' pointer's _data field to contain junk after the
+// lock. We require the caller to take the lock before making the ProfileData
+// structure. Currently the only caller is InterpreterRuntime::update_mdp_for_ret
+address RetData::fixup_ret(int return_bci, MethodData* h_mdo) {
+ // First find the mdp which corresponds to the return bci.
+ address mdp = h_mdo->bci_to_dp(return_bci);
+
+ // Now check to see if any of the cache slots are open.
+ for (uint row = 0; row < row_limit(); row++) {
+ if (bci(row) == no_bci) {
+ set_bci_displacement(row, mdp - dp());
+ set_bci_count(row, DataLayout::counter_increment);
+ // Barrier to ensure displacement is written before the bci; allows
+ // the interpreter to read displacement without fear of race condition.
+ release_set_bci(row, return_bci);
+ break;
+ }
+ }
+ return mdp;
+}
+
+#ifdef CC_INTERP
+DataLayout* RetData::advance(MethodData *md, int bci) {
+ return (DataLayout*) md->bci_to_dp(bci);
+}
+#endif // CC_INTERP
+
+void RetData::print_data_on(outputStream* st, const char* extra) const {
+ print_shared(st, "RetData", extra);
+ uint row;
+ int entries = 0;
+ for (row = 0; row < row_limit(); row++) {
+ if (bci(row) != no_bci) entries++;
+ }
+ st->print_cr("count(%u) entries(%u)", count(), entries);
+ for (row = 0; row < row_limit(); row++) {
+ if (bci(row) != no_bci) {
+ tab(st);
+ st->print_cr("bci(%d: count(%u) displacement(%d))",
+ bci(row), bci_count(row), bci_displacement(row));
+ }
+ }
+}
+
+// ==================================================================
+// BranchData
+//
+// A BranchData is used to access profiling data for a two-way branch.
+// It consists of taken and not_taken counts as well as a data displacement
+// for the taken case.
+
+void BranchData::post_initialize(BytecodeStream* stream, MethodData* mdo) {
+ assert(stream->bci() == bci(), "wrong pos");
+ int target = stream->dest();
+ int my_di = mdo->dp_to_di(dp());
+ int target_di = mdo->bci_to_di(target);
+ int offset = target_di - my_di;
+ set_displacement(offset);
+}
+
+void BranchData::print_data_on(outputStream* st, const char* extra) const {
+ print_shared(st, "BranchData", extra);
+ st->print_cr("taken(%u) displacement(%d)",
+ taken(), displacement());
+ tab(st);
+ st->print_cr("not taken(%u)", not_taken());
+}
+
+// ==================================================================
+// MultiBranchData
+//
+// A MultiBranchData is used to access profiling information for
+// a multi-way branch (*switch bytecodes). It consists of a series
+// of (count, displacement) pairs, which count the number of times each
+// case was taken and specify the data displacment for each branch target.
+
+int MultiBranchData::compute_cell_count(BytecodeStream* stream) {
+ int cell_count = 0;
+ if (stream->code() == Bytecodes::_tableswitch) {
+ Bytecode_tableswitch sw(stream->method()(), stream->bcp());
+ cell_count = 1 + per_case_cell_count * (1 + sw.length()); // 1 for default
+ } else {
+ Bytecode_lookupswitch sw(stream->method()(), stream->bcp());
+ cell_count = 1 + per_case_cell_count * (sw.number_of_pairs() + 1); // 1 for default
+ }
+ return cell_count;
+}
+
+void MultiBranchData::post_initialize(BytecodeStream* stream,
+ MethodData* mdo) {
+ assert(stream->bci() == bci(), "wrong pos");
+ int target;
+ int my_di;
+ int target_di;
+ int offset;
+ if (stream->code() == Bytecodes::_tableswitch) {
+ Bytecode_tableswitch sw(stream->method()(), stream->bcp());
+ int len = sw.length();
+ assert(array_len() == per_case_cell_count * (len + 1), "wrong len");
+ for (int count = 0; count < len; count++) {
+ target = sw.dest_offset_at(count) + bci();
+ my_di = mdo->dp_to_di(dp());
+ target_di = mdo->bci_to_di(target);
+ offset = target_di - my_di;
+ set_displacement_at(count, offset);
+ }
+ target = sw.default_offset() + bci();
+ my_di = mdo->dp_to_di(dp());
+ target_di = mdo->bci_to_di(target);
+ offset = target_di - my_di;
+ set_default_displacement(offset);
+
+ } else {
+ Bytecode_lookupswitch sw(stream->method()(), stream->bcp());
+ int npairs = sw.number_of_pairs();
+ assert(array_len() == per_case_cell_count * (npairs + 1), "wrong len");
+ for (int count = 0; count < npairs; count++) {
+ LookupswitchPair pair = sw.pair_at(count);
+ target = pair.offset() + bci();
+ my_di = mdo->dp_to_di(dp());
+ target_di = mdo->bci_to_di(target);
+ offset = target_di - my_di;
+ set_displacement_at(count, offset);
+ }
+ target = sw.default_offset() + bci();
+ my_di = mdo->dp_to_di(dp());
+ target_di = mdo->bci_to_di(target);
+ offset = target_di - my_di;
+ set_default_displacement(offset);
+ }
+}
+
+void MultiBranchData::print_data_on(outputStream* st, const char* extra) const {
+ print_shared(st, "MultiBranchData", extra);
+ st->print_cr("default_count(%u) displacement(%d)",
+ default_count(), default_displacement());
+ int cases = number_of_cases();
+ for (int i = 0; i < cases; i++) {
+ tab(st);
+ st->print_cr("count(%u) displacement(%d)",
+ count_at(i), displacement_at(i));
+ }
+}
+
+void ArgInfoData::print_data_on(outputStream* st, const char* extra) const {
+ print_shared(st, "ArgInfoData", extra);
+ int nargs = number_of_args();
+ for (int i = 0; i < nargs; i++) {
+ st->print(" 0x%x", arg_modified(i));
+ }
+ st->cr();
+}
+
+int ParametersTypeData::compute_cell_count(Method* m) {
+ if (!MethodData::profile_parameters_for_method(m)) {
+ return 0;
+ }
+ int max = TypeProfileParmsLimit == -1 ? INT_MAX : TypeProfileParmsLimit;
+ int obj_args = TypeStackSlotEntries::compute_cell_count(m->signature(), !m->is_static(), max);
+ if (obj_args > 0) {
+ return obj_args + 1; // 1 cell for array len
+ }
+ return 0;
+}
+
+void ParametersTypeData::post_initialize(BytecodeStream* stream, MethodData* mdo) {
+ _parameters.post_initialize(mdo->method()->signature(), !mdo->method()->is_static(), true);
+}
+
+bool ParametersTypeData::profiling_enabled() {
+ return MethodData::profile_parameters();
+}
+
+void ParametersTypeData::print_data_on(outputStream* st, const char* extra) const {
+ st->print("parameter types"); // FIXME extra ignored?
+ _parameters.print_data_on(st);
+}
+
+void SpeculativeTrapData::print_data_on(outputStream* st, const char* extra) const {
+ print_shared(st, "SpeculativeTrapData", extra);
+ tab(st);
+ method()->print_short_name(st);
+ st->cr();
+}
+
+// ==================================================================
+// MethodData*
+//
+// A MethodData* holds information which has been collected about
+// a method.
+
+MethodData* MethodData::allocate(ClassLoaderData* loader_data, const methodHandle& method, TRAPS) {
+ int size = MethodData::compute_allocation_size_in_words(method);
+
+ return new (loader_data, size, MetaspaceObj::MethodDataType, THREAD)
+ MethodData(method(), size, THREAD);
+}
+
+int MethodData::bytecode_cell_count(Bytecodes::Code code) {
+ if (is_client_compilation_mode_vm()) {
+ return no_profile_data;
+ }
+ switch (code) {
+ case Bytecodes::_checkcast:
+ case Bytecodes::_instanceof:
+ case Bytecodes::_aastore:
+ if (TypeProfileCasts) {
+ return ReceiverTypeData::static_cell_count();
+ } else {
+ return BitData::static_cell_count();
+ }
+ case Bytecodes::_invokespecial:
+ case Bytecodes::_invokestatic:
+ if (MethodData::profile_arguments() || MethodData::profile_return()) {
+ return variable_cell_count;
+ } else {
+ return CounterData::static_cell_count();
+ }
+ case Bytecodes::_goto:
+ case Bytecodes::_goto_w:
+ case Bytecodes::_jsr:
+ case Bytecodes::_jsr_w:
+ return JumpData::static_cell_count();
+ case Bytecodes::_invokevirtual:
+ case Bytecodes::_invokeinterface:
+ if (MethodData::profile_arguments() || MethodData::profile_return()) {
+ return variable_cell_count;
+ } else {
+ return VirtualCallData::static_cell_count();
+ }
+ case Bytecodes::_invokedynamic:
+ if (MethodData::profile_arguments() || MethodData::profile_return()) {
+ return variable_cell_count;
+ } else {
+ return CounterData::static_cell_count();
+ }
+ case Bytecodes::_ret:
+ return RetData::static_cell_count();
+ case Bytecodes::_ifeq:
+ case Bytecodes::_ifne:
+ case Bytecodes::_iflt:
+ case Bytecodes::_ifge:
+ case Bytecodes::_ifgt:
+ case Bytecodes::_ifle:
+ case Bytecodes::_if_icmpeq:
+ case Bytecodes::_if_icmpne:
+ case Bytecodes::_if_icmplt:
+ case Bytecodes::_if_icmpge:
+ case Bytecodes::_if_icmpgt:
+ case Bytecodes::_if_icmple:
+ case Bytecodes::_if_acmpeq:
+ case Bytecodes::_if_acmpne:
+ case Bytecodes::_ifnull:
+ case Bytecodes::_ifnonnull:
+ return BranchData::static_cell_count();
+ case Bytecodes::_lookupswitch:
+ case Bytecodes::_tableswitch:
+ return variable_cell_count;
+ default:
+ return no_profile_data;
+ }
+}
+
+// Compute the size of the profiling information corresponding to
+// the current bytecode.
+int MethodData::compute_data_size(BytecodeStream* stream) {
+ int cell_count = bytecode_cell_count(stream->code());
+ if (cell_count == no_profile_data) {
+ return 0;
+ }
+ if (cell_count == variable_cell_count) {
+ switch (stream->code()) {
+ case Bytecodes::_lookupswitch:
+ case Bytecodes::_tableswitch:
+ cell_count = MultiBranchData::compute_cell_count(stream);
+ break;
+ case Bytecodes::_invokespecial:
+ case Bytecodes::_invokestatic:
+ case Bytecodes::_invokedynamic:
+ assert(MethodData::profile_arguments() || MethodData::profile_return(), "should be collecting args profile");
+ if (profile_arguments_for_invoke(stream->method(), stream->bci()) ||
+ profile_return_for_invoke(stream->method(), stream->bci())) {
+ cell_count = CallTypeData::compute_cell_count(stream);
+ } else {
+ cell_count = CounterData::static_cell_count();
+ }
+ break;
+ case Bytecodes::_invokevirtual:
+ case Bytecodes::_invokeinterface: {
+ assert(MethodData::profile_arguments() || MethodData::profile_return(), "should be collecting args profile");
+ if (profile_arguments_for_invoke(stream->method(), stream->bci()) ||
+ profile_return_for_invoke(stream->method(), stream->bci())) {
+ cell_count = VirtualCallTypeData::compute_cell_count(stream);
+ } else {
+ cell_count = VirtualCallData::static_cell_count();
+ }
+ break;
+ }
+ default:
+ fatal("unexpected bytecode for var length profile data");
+ }
+ }
+ // Note: cell_count might be zero, meaning that there is just
+ // a DataLayout header, with no extra cells.
+ assert(cell_count >= 0, "sanity");
+ return DataLayout::compute_size_in_bytes(cell_count);
+}
+
+bool MethodData::is_speculative_trap_bytecode(Bytecodes::Code code) {
+ // Bytecodes for which we may use speculation
+ switch (code) {
+ case Bytecodes::_checkcast:
+ case Bytecodes::_instanceof:
+ case Bytecodes::_aastore:
+ case Bytecodes::_invokevirtual:
+ case Bytecodes::_invokeinterface:
+ case Bytecodes::_if_acmpeq:
+ case Bytecodes::_if_acmpne:
+ case Bytecodes::_ifnull:
+ case Bytecodes::_ifnonnull:
+ case Bytecodes::_invokestatic:
+#ifdef COMPILER2
+ if (is_server_compilation_mode_vm()) {
+ return UseTypeSpeculation;
+ }
+#endif
+ default:
+ return false;
+ }
+ return false;
+}
+
+int MethodData::compute_extra_data_count(int data_size, int empty_bc_count, bool needs_speculative_traps) {
+#if INCLUDE_JVMCI
+ if (ProfileTraps) {
+ // Assume that up to 30% of the possibly trapping BCIs with no MDP will need to allocate one.
+ int extra_data_count = MIN2(empty_bc_count, MAX2(4, (empty_bc_count * 30) / 100));
+
+ // Make sure we have a minimum number of extra data slots to
+ // allocate SpeculativeTrapData entries. We would want to have one
+ // entry per compilation that inlines this method and for which
+ // some type speculation assumption fails. So the room we need for
+ // the SpeculativeTrapData entries doesn't directly depend on the
+ // size of the method. Because it's hard to estimate, we reserve
+ // space for an arbitrary number of entries.
+ int spec_data_count = (needs_speculative_traps ? SpecTrapLimitExtraEntries : 0) *
+ (SpeculativeTrapData::static_cell_count() + DataLayout::header_size_in_cells());
+
+ return MAX2(extra_data_count, spec_data_count);
+ } else {
+ return 0;
+ }
+#else // INCLUDE_JVMCI
+ if (ProfileTraps) {
+ // Assume that up to 3% of BCIs with no MDP will need to allocate one.
+ int extra_data_count = (uint)(empty_bc_count * 3) / 128 + 1;
+ // If the method is large, let the extra BCIs grow numerous (to ~1%).
+ int one_percent_of_data
+ = (uint)data_size / (DataLayout::header_size_in_bytes()*128);
+ if (extra_data_count < one_percent_of_data)
+ extra_data_count = one_percent_of_data;
+ if (extra_data_count > empty_bc_count)
+ extra_data_count = empty_bc_count; // no need for more
+
+ // Make sure we have a minimum number of extra data slots to
+ // allocate SpeculativeTrapData entries. We would want to have one
+ // entry per compilation that inlines this method and for which
+ // some type speculation assumption fails. So the room we need for
+ // the SpeculativeTrapData entries doesn't directly depend on the
+ // size of the method. Because it's hard to estimate, we reserve
+ // space for an arbitrary number of entries.
+ int spec_data_count = (needs_speculative_traps ? SpecTrapLimitExtraEntries : 0) *
+ (SpeculativeTrapData::static_cell_count() + DataLayout::header_size_in_cells());
+
+ return MAX2(extra_data_count, spec_data_count);
+ } else {
+ return 0;
+ }
+#endif // INCLUDE_JVMCI
+}
+
+// Compute the size of the MethodData* necessary to store
+// profiling information about a given method. Size is in bytes.
+int MethodData::compute_allocation_size_in_bytes(const methodHandle& method) {
+ int data_size = 0;
+ BytecodeStream stream(method);
+ Bytecodes::Code c;
+ int empty_bc_count = 0; // number of bytecodes lacking data
+ bool needs_speculative_traps = false;
+ while ((c = stream.next()) >= 0) {
+ int size_in_bytes = compute_data_size(&stream);
+ data_size += size_in_bytes;
+ if (size_in_bytes == 0 JVMCI_ONLY(&& Bytecodes::can_trap(c))) empty_bc_count += 1;
+ needs_speculative_traps = needs_speculative_traps || is_speculative_trap_bytecode(c);
+ }
+ int object_size = in_bytes(data_offset()) + data_size;
+
+ // Add some extra DataLayout cells (at least one) to track stray traps.
+ int extra_data_count = compute_extra_data_count(data_size, empty_bc_count, needs_speculative_traps);
+ object_size += extra_data_count * DataLayout::compute_size_in_bytes(0);
+
+ // Add a cell to record information about modified arguments.
+ int arg_size = method->size_of_parameters();
+ object_size += DataLayout::compute_size_in_bytes(arg_size+1);
+
+ // Reserve room for an area of the MDO dedicated to profiling of
+ // parameters
+ int args_cell = ParametersTypeData::compute_cell_count(method());
+ if (args_cell > 0) {
+ object_size += DataLayout::compute_size_in_bytes(args_cell);
+ }
+ return object_size;
+}
+
+// Compute the size of the MethodData* necessary to store
+// profiling information about a given method. Size is in words
+int MethodData::compute_allocation_size_in_words(const methodHandle& method) {
+ int byte_size = compute_allocation_size_in_bytes(method);
+ int word_size = align_up(byte_size, BytesPerWord) / BytesPerWord;
+ return align_metadata_size(word_size);
+}
+
+// Initialize an individual data segment. Returns the size of
+// the segment in bytes.
+int MethodData::initialize_data(BytecodeStream* stream,
+ int data_index) {
+ if (is_client_compilation_mode_vm()) {
+ return 0;
+ }
+ int cell_count = -1;
+ int tag = DataLayout::no_tag;
+ DataLayout* data_layout = data_layout_at(data_index);
+ Bytecodes::Code c = stream->code();
+ switch (c) {
+ case Bytecodes::_checkcast:
+ case Bytecodes::_instanceof:
+ case Bytecodes::_aastore:
+ if (TypeProfileCasts) {
+ cell_count = ReceiverTypeData::static_cell_count();
+ tag = DataLayout::receiver_type_data_tag;
+ } else {
+ cell_count = BitData::static_cell_count();
+ tag = DataLayout::bit_data_tag;
+ }
+ break;
+ case Bytecodes::_invokespecial:
+ case Bytecodes::_invokestatic: {
+ int counter_data_cell_count = CounterData::static_cell_count();
+ if (profile_arguments_for_invoke(stream->method(), stream->bci()) ||
+ profile_return_for_invoke(stream->method(), stream->bci())) {
+ cell_count = CallTypeData::compute_cell_count(stream);
+ } else {
+ cell_count = counter_data_cell_count;
+ }
+ if (cell_count > counter_data_cell_count) {
+ tag = DataLayout::call_type_data_tag;
+ } else {
+ tag = DataLayout::counter_data_tag;
+ }
+ break;
+ }
+ case Bytecodes::_goto:
+ case Bytecodes::_goto_w:
+ case Bytecodes::_jsr:
+ case Bytecodes::_jsr_w:
+ cell_count = JumpData::static_cell_count();
+ tag = DataLayout::jump_data_tag;
+ break;
+ case Bytecodes::_invokevirtual:
+ case Bytecodes::_invokeinterface: {
+ int virtual_call_data_cell_count = VirtualCallData::static_cell_count();
+ if (profile_arguments_for_invoke(stream->method(), stream->bci()) ||
+ profile_return_for_invoke(stream->method(), stream->bci())) {
+ cell_count = VirtualCallTypeData::compute_cell_count(stream);
+ } else {
+ cell_count = virtual_call_data_cell_count;
+ }
+ if (cell_count > virtual_call_data_cell_count) {
+ tag = DataLayout::virtual_call_type_data_tag;
+ } else {
+ tag = DataLayout::virtual_call_data_tag;
+ }
+ break;
+ }
+ case Bytecodes::_invokedynamic: {
+ // %%% should make a type profile for any invokedynamic that takes a ref argument
+ int counter_data_cell_count = CounterData::static_cell_count();
+ if (profile_arguments_for_invoke(stream->method(), stream->bci()) ||
+ profile_return_for_invoke(stream->method(), stream->bci())) {
+ cell_count = CallTypeData::compute_cell_count(stream);
+ } else {
+ cell_count = counter_data_cell_count;
+ }
+ if (cell_count > counter_data_cell_count) {
+ tag = DataLayout::call_type_data_tag;
+ } else {
+ tag = DataLayout::counter_data_tag;
+ }
+ break;
+ }
+ case Bytecodes::_ret:
+ cell_count = RetData::static_cell_count();
+ tag = DataLayout::ret_data_tag;
+ break;
+ case Bytecodes::_ifeq:
+ case Bytecodes::_ifne:
+ case Bytecodes::_iflt:
+ case Bytecodes::_ifge:
+ case Bytecodes::_ifgt:
+ case Bytecodes::_ifle:
+ case Bytecodes::_if_icmpeq:
+ case Bytecodes::_if_icmpne:
+ case Bytecodes::_if_icmplt:
+ case Bytecodes::_if_icmpge:
+ case Bytecodes::_if_icmpgt:
+ case Bytecodes::_if_icmple:
+ case Bytecodes::_if_acmpeq:
+ case Bytecodes::_if_acmpne:
+ case Bytecodes::_ifnull:
+ case Bytecodes::_ifnonnull:
+ cell_count = BranchData::static_cell_count();
+ tag = DataLayout::branch_data_tag;
+ break;
+ case Bytecodes::_lookupswitch:
+ case Bytecodes::_tableswitch:
+ cell_count = MultiBranchData::compute_cell_count(stream);
+ tag = DataLayout::multi_branch_data_tag;
+ break;
+ default:
+ break;
+ }
+ assert(tag == DataLayout::multi_branch_data_tag ||
+ ((MethodData::profile_arguments() || MethodData::profile_return()) &&
+ (tag == DataLayout::call_type_data_tag ||
+ tag == DataLayout::counter_data_tag ||
+ tag == DataLayout::virtual_call_type_data_tag ||
+ tag == DataLayout::virtual_call_data_tag)) ||
+ cell_count == bytecode_cell_count(c), "cell counts must agree");
+ if (cell_count >= 0) {
+ assert(tag != DataLayout::no_tag, "bad tag");
+ assert(bytecode_has_profile(c), "agree w/ BHP");
+ data_layout->initialize(tag, stream->bci(), cell_count);
+ return DataLayout::compute_size_in_bytes(cell_count);
+ } else {
+ assert(!bytecode_has_profile(c), "agree w/ !BHP");
+ return 0;
+ }
+}
+
+// Get the data at an arbitrary (sort of) data index.
+ProfileData* MethodData::data_at(int data_index) const {
+ if (out_of_bounds(data_index)) {
+ return NULL;
+ }
+ DataLayout* data_layout = data_layout_at(data_index);
+ return data_layout->data_in();
+}
+
+ProfileData* DataLayout::data_in() {
+ switch (tag()) {
+ case DataLayout::no_tag:
+ default:
+ ShouldNotReachHere();
+ return NULL;
+ case DataLayout::bit_data_tag:
+ return new BitData(this);
+ case DataLayout::counter_data_tag:
+ return new CounterData(this);
+ case DataLayout::jump_data_tag:
+ return new JumpData(this);
+ case DataLayout::receiver_type_data_tag:
+ return new ReceiverTypeData(this);
+ case DataLayout::virtual_call_data_tag:
+ return new VirtualCallData(this);
+ case DataLayout::ret_data_tag:
+ return new RetData(this);
+ case DataLayout::branch_data_tag:
+ return new BranchData(this);
+ case DataLayout::multi_branch_data_tag:
+ return new MultiBranchData(this);
+ case DataLayout::arg_info_data_tag:
+ return new ArgInfoData(this);
+ case DataLayout::call_type_data_tag:
+ return new CallTypeData(this);
+ case DataLayout::virtual_call_type_data_tag:
+ return new VirtualCallTypeData(this);
+ case DataLayout::parameters_type_data_tag:
+ return new ParametersTypeData(this);
+ case DataLayout::speculative_trap_data_tag:
+ return new SpeculativeTrapData(this);
+ }
+}
+
+// Iteration over data.
+ProfileData* MethodData::next_data(ProfileData* current) const {
+ int current_index = dp_to_di(current->dp());
+ int next_index = current_index + current->size_in_bytes();
+ ProfileData* next = data_at(next_index);
+ return next;
+}
+
+// Give each of the data entries a chance to perform specific
+// data initialization.
+void MethodData::post_initialize(BytecodeStream* stream) {
+ ResourceMark rm;
+ ProfileData* data;
+ for (data = first_data(); is_valid(data); data = next_data(data)) {
+ stream->set_start(data->bci());
+ stream->next();
+ data->post_initialize(stream, this);
+ }
+ if (_parameters_type_data_di != no_parameters) {
+ parameters_type_data()->post_initialize(NULL, this);
+ }
+}
+
+// Initialize the MethodData* corresponding to a given method.
+MethodData::MethodData(const methodHandle& method, int size, TRAPS)
+ : _extra_data_lock(Monitor::leaf, "MDO extra data lock"),
+ _parameters_type_data_di(parameters_uninitialized) {
+ // Set the method back-pointer.
+ _method = method();
+ initialize();
+}
+
+void MethodData::initialize() {
+ NoSafepointVerifier no_safepoint; // init function atomic wrt GC
+ ResourceMark rm;
+
+ init();
+ set_creation_mileage(mileage_of(method()));
+
+ // Go through the bytecodes and allocate and initialize the
+ // corresponding data cells.
+ int data_size = 0;
+ int empty_bc_count = 0; // number of bytecodes lacking data
+ _data[0] = 0; // apparently not set below.
+ BytecodeStream stream(method());
+ Bytecodes::Code c;
+ bool needs_speculative_traps = false;
+ while ((c = stream.next()) >= 0) {
+ int size_in_bytes = initialize_data(&stream, data_size);
+ data_size += size_in_bytes;
+ if (size_in_bytes == 0 JVMCI_ONLY(&& Bytecodes::can_trap(c))) empty_bc_count += 1;
+ needs_speculative_traps = needs_speculative_traps || is_speculative_trap_bytecode(c);
+ }
+ _data_size = data_size;
+ int object_size = in_bytes(data_offset()) + data_size;
+
+ // Add some extra DataLayout cells (at least one) to track stray traps.
+ int extra_data_count = compute_extra_data_count(data_size, empty_bc_count, needs_speculative_traps);
+ int extra_size = extra_data_count * DataLayout::compute_size_in_bytes(0);
+
+ // Let's zero the space for the extra data
+ Copy::zero_to_bytes(((address)_data) + data_size, extra_size);
+
+ // Add a cell to record information about modified arguments.
+ // Set up _args_modified array after traps cells so that
+ // the code for traps cells works.
+ DataLayout *dp = data_layout_at(data_size + extra_size);
+
+ int arg_size = method()->size_of_parameters();
+ dp->initialize(DataLayout::arg_info_data_tag, 0, arg_size+1);
+
+ int arg_data_size = DataLayout::compute_size_in_bytes(arg_size+1);
+ object_size += extra_size + arg_data_size;
+
+ int parms_cell = ParametersTypeData::compute_cell_count(method());
+ // If we are profiling parameters, we reserver an area near the end
+ // of the MDO after the slots for bytecodes (because there's no bci
+ // for method entry so they don't fit with the framework for the
+ // profiling of bytecodes). We store the offset within the MDO of
+ // this area (or -1 if no parameter is profiled)
+ if (parms_cell > 0) {
+ object_size += DataLayout::compute_size_in_bytes(parms_cell);
+ _parameters_type_data_di = data_size + extra_size + arg_data_size;
+ DataLayout *dp = data_layout_at(data_size + extra_size + arg_data_size);
+ dp->initialize(DataLayout::parameters_type_data_tag, 0, parms_cell);
+ } else {
+ _parameters_type_data_di = no_parameters;
+ }
+
+ // Set an initial hint. Don't use set_hint_di() because
+ // first_di() may be out of bounds if data_size is 0.
+ // In that situation, _hint_di is never used, but at
+ // least well-defined.
+ _hint_di = first_di();
+
+ post_initialize(&stream);
+
+ assert(object_size == compute_allocation_size_in_bytes(methodHandle(_method)), "MethodData: computed size != initialized size");
+ set_size(object_size);
+}
+
+void MethodData::init() {
+ _invocation_counter.init();
+ _backedge_counter.init();
+ _invocation_counter_start = 0;
+ _backedge_counter_start = 0;
+
+ // Set per-method invoke- and backedge mask.
+ double scale = 1.0;
+ CompilerOracle::has_option_value(_method, "CompileThresholdScaling", scale);
+ _invoke_mask = right_n_bits(Arguments::scaled_freq_log(Tier0InvokeNotifyFreqLog, scale)) << InvocationCounter::count_shift;
+ _backedge_mask = right_n_bits(Arguments::scaled_freq_log(Tier0BackedgeNotifyFreqLog, scale)) << InvocationCounter::count_shift;
+
+ _tenure_traps = 0;
+ _num_loops = 0;
+ _num_blocks = 0;
+ _would_profile = unknown;
+
+#if INCLUDE_JVMCI
+ _jvmci_ir_size = 0;
+#endif
+
+#if INCLUDE_RTM_OPT
+ _rtm_state = NoRTM; // No RTM lock eliding by default
+ if (UseRTMLocking &&
+ !CompilerOracle::has_option_string(_method, "NoRTMLockEliding")) {
+ if (CompilerOracle::has_option_string(_method, "UseRTMLockEliding") || !UseRTMDeopt) {
+ // Generate RTM lock eliding code without abort ratio calculation code.
+ _rtm_state = UseRTM;
+ } else if (UseRTMDeopt) {
+ // Generate RTM lock eliding code and include abort ratio calculation
+ // code if UseRTMDeopt is on.
+ _rtm_state = ProfileRTM;
+ }
+ }
+#endif
+
+ // Initialize flags and trap history.
+ _nof_decompiles = 0;
+ _nof_overflow_recompiles = 0;
+ _nof_overflow_traps = 0;
+ clear_escape_info();
+ assert(sizeof(_trap_hist) % sizeof(HeapWord) == 0, "align");
+ Copy::zero_to_words((HeapWord*) &_trap_hist,
+ sizeof(_trap_hist) / sizeof(HeapWord));
+}
+
+// Get a measure of how much mileage the method has on it.
+int MethodData::mileage_of(Method* method) {
+ int mileage = 0;
+ if (TieredCompilation) {
+ mileage = MAX2(method->invocation_count(), method->backedge_count());
+ } else {
+ int iic = method->interpreter_invocation_count();
+ if (mileage < iic) mileage = iic;
+ MethodCounters* mcs = method->method_counters();
+ if (mcs != NULL) {
+ InvocationCounter* ic = mcs->invocation_counter();
+ InvocationCounter* bc = mcs->backedge_counter();
+ int icval = ic->count();
+ if (ic->carry()) icval += CompileThreshold;
+ if (mileage < icval) mileage = icval;
+ int bcval = bc->count();
+ if (bc->carry()) bcval += CompileThreshold;
+ if (mileage < bcval) mileage = bcval;
+ }
+ }
+ return mileage;
+}
+
+bool MethodData::is_mature() const {
+ return CompilationPolicy::policy()->is_mature(_method);
+}
+
+// Translate a bci to its corresponding data index (di).
+address MethodData::bci_to_dp(int bci) {
+ ResourceMark rm;
+ ProfileData* data = data_before(bci);
+ ProfileData* prev = NULL;
+ for ( ; is_valid(data); data = next_data(data)) {
+ if (data->bci() >= bci) {
+ if (data->bci() == bci) set_hint_di(dp_to_di(data->dp()));
+ else if (prev != NULL) set_hint_di(dp_to_di(prev->dp()));
+ return data->dp();
+ }
+ prev = data;
+ }
+ return (address)limit_data_position();
+}
+
+// Translate a bci to its corresponding data, or NULL.
+ProfileData* MethodData::bci_to_data(int bci) {
+ ProfileData* data = data_before(bci);
+ for ( ; is_valid(data); data = next_data(data)) {
+ if (data->bci() == bci) {
+ set_hint_di(dp_to_di(data->dp()));
+ return data;
+ } else if (data->bci() > bci) {
+ break;
+ }
+ }
+ return bci_to_extra_data(bci, NULL, false);
+}
+
+DataLayout* MethodData::next_extra(DataLayout* dp) {
+ int nb_cells = 0;
+ switch(dp->tag()) {
+ case DataLayout::bit_data_tag:
+ case DataLayout::no_tag:
+ nb_cells = BitData::static_cell_count();
+ break;
+ case DataLayout::speculative_trap_data_tag:
+ nb_cells = SpeculativeTrapData::static_cell_count();
+ break;
+ default:
+ fatal("unexpected tag %d", dp->tag());
+ }
+ return (DataLayout*)((address)dp + DataLayout::compute_size_in_bytes(nb_cells));
+}
+
+ProfileData* MethodData::bci_to_extra_data_helper(int bci, Method* m, DataLayout*& dp, bool concurrent) {
+ DataLayout* end = args_data_limit();
+
+ for (;; dp = next_extra(dp)) {
+ assert(dp < end, "moved past end of extra data");
+ // No need for "OrderAccess::load_acquire" ops,
+ // since the data structure is monotonic.
+ switch(dp->tag()) {
+ case DataLayout::no_tag:
+ return NULL;
+ case DataLayout::arg_info_data_tag:
+ dp = end;
+ return NULL; // ArgInfoData is at the end of extra data section.
+ case DataLayout::bit_data_tag:
+ if (m == NULL && dp->bci() == bci) {
+ return new BitData(dp);
+ }
+ break;
+ case DataLayout::speculative_trap_data_tag:
+ if (m != NULL) {
+ SpeculativeTrapData* data = new SpeculativeTrapData(dp);
+ // data->method() may be null in case of a concurrent
+ // allocation. Maybe it's for the same method. Try to use that
+ // entry in that case.
+ if (dp->bci() == bci) {
+ if (data->method() == NULL) {
+ assert(concurrent, "impossible because no concurrent allocation");
+ return NULL;
+ } else if (data->method() == m) {
+ return data;
+ }
+ }
+ }
+ break;
+ default:
+ fatal("unexpected tag %d", dp->tag());
+ }
+ }
+ return NULL;
+}
+
+
+// Translate a bci to its corresponding extra data, or NULL.
+ProfileData* MethodData::bci_to_extra_data(int bci, Method* m, bool create_if_missing) {
+ // This code assumes an entry for a SpeculativeTrapData is 2 cells
+ assert(2*DataLayout::compute_size_in_bytes(BitData::static_cell_count()) ==
+ DataLayout::compute_size_in_bytes(SpeculativeTrapData::static_cell_count()),
+ "code needs to be adjusted");
+
+ // Do not create one of these if method has been redefined.
+ if (m != NULL && m->is_old()) {
+ return NULL;
+ }
+
+ DataLayout* dp = extra_data_base();
+ DataLayout* end = args_data_limit();
+
+ // Allocation in the extra data space has to be atomic because not
+ // all entries have the same size and non atomic concurrent
+ // allocation would result in a corrupted extra data space.
+ ProfileData* result = bci_to_extra_data_helper(bci, m, dp, true);
+ if (result != NULL) {
+ return result;
+ }
+
+ if (create_if_missing && dp < end) {
+ MutexLocker ml(&_extra_data_lock);
+ // Check again now that we have the lock. Another thread may
+ // have added extra data entries.
+ ProfileData* result = bci_to_extra_data_helper(bci, m, dp, false);
+ if (result != NULL || dp >= end) {
+ return result;
+ }
+
+ assert(dp->tag() == DataLayout::no_tag || (dp->tag() == DataLayout::speculative_trap_data_tag && m != NULL), "should be free");
+ assert(next_extra(dp)->tag() == DataLayout::no_tag || next_extra(dp)->tag() == DataLayout::arg_info_data_tag, "should be free or arg info");
+ u1 tag = m == NULL ? DataLayout::bit_data_tag : DataLayout::speculative_trap_data_tag;
+ // SpeculativeTrapData is 2 slots. Make sure we have room.
+ if (m != NULL && next_extra(dp)->tag() != DataLayout::no_tag) {
+ return NULL;
+ }
+ DataLayout temp;
+ temp.initialize(tag, bci, 0);
+
+ dp->set_header(temp.header());
+ assert(dp->tag() == tag, "sane");
+ assert(dp->bci() == bci, "no concurrent allocation");
+ if (tag == DataLayout::bit_data_tag) {
+ return new BitData(dp);
+ } else {
+ SpeculativeTrapData* data = new SpeculativeTrapData(dp);
+ data->set_method(m);
+ return data;
+ }
+ }
+ return NULL;
+}
+
+ArgInfoData *MethodData::arg_info() {
+ DataLayout* dp = extra_data_base();
+ DataLayout* end = args_data_limit();
+ for (; dp < end; dp = next_extra(dp)) {
+ if (dp->tag() == DataLayout::arg_info_data_tag)
+ return new ArgInfoData(dp);
+ }
+ return NULL;
+}
+
+// Printing
+
+void MethodData::print_on(outputStream* st) const {
+ assert(is_methodData(), "should be method data");
+ st->print("method data for ");
+ method()->print_value_on(st);
+ st->cr();
+ print_data_on(st);
+}
+
+void MethodData::print_value_on(outputStream* st) const {
+ assert(is_methodData(), "should be method data");
+ st->print("method data for ");
+ method()->print_value_on(st);
+}
+
+void MethodData::print_data_on(outputStream* st) const {
+ ResourceMark rm;
+ ProfileData* data = first_data();
+ if (_parameters_type_data_di != no_parameters) {
+ parameters_type_data()->print_data_on(st);
+ }
+ for ( ; is_valid(data); data = next_data(data)) {
+ st->print("%d", dp_to_di(data->dp()));
+ st->fill_to(6);
+ data->print_data_on(st, this);
+ }
+ st->print_cr("--- Extra data:");
+ DataLayout* dp = extra_data_base();
+ DataLayout* end = args_data_limit();
+ for (;; dp = next_extra(dp)) {
+ assert(dp < end, "moved past end of extra data");
+ // No need for "OrderAccess::load_acquire" ops,
+ // since the data structure is monotonic.
+ switch(dp->tag()) {
+ case DataLayout::no_tag:
+ continue;
+ case DataLayout::bit_data_tag:
+ data = new BitData(dp);
+ break;
+ case DataLayout::speculative_trap_data_tag:
+ data = new SpeculativeTrapData(dp);
+ break;
+ case DataLayout::arg_info_data_tag:
+ data = new ArgInfoData(dp);
+ dp = end; // ArgInfoData is at the end of extra data section.
+ break;
+ default:
+ fatal("unexpected tag %d", dp->tag());
+ }
+ st->print("%d", dp_to_di(data->dp()));
+ st->fill_to(6);
+ data->print_data_on(st);
+ if (dp >= end) return;
+ }
+}
+
+#if INCLUDE_SERVICES
+// Size Statistics
+void MethodData::collect_statistics(KlassSizeStats *sz) const {
+ int n = sz->count(this);
+ sz->_method_data_bytes += n;
+ sz->_method_all_bytes += n;
+ sz->_rw_bytes += n;
+}
+#endif // INCLUDE_SERVICES
+
+// Verification
+
+void MethodData::verify_on(outputStream* st) {
+ guarantee(is_methodData(), "object must be method data");
+ // guarantee(m->is_perm(), "should be in permspace");
+ this->verify_data_on(st);
+}
+
+void MethodData::verify_data_on(outputStream* st) {
+ NEEDS_CLEANUP;
+ // not yet implemented.
+}
+
+bool MethodData::profile_jsr292(const methodHandle& m, int bci) {
+ if (m->is_compiled_lambda_form()) {
+ return true;
+ }
+
+ Bytecode_invoke inv(m , bci);
+ return inv.is_invokedynamic() || inv.is_invokehandle();
+}
+
+bool MethodData::profile_unsafe(const methodHandle& m, int bci) {
+ Bytecode_invoke inv(m , bci);
+ if (inv.is_invokevirtual() && inv.klass() == vmSymbols::jdk_internal_misc_Unsafe()) {
+ ResourceMark rm;
+ char* name = inv.name()->as_C_string();
+ if (!strncmp(name, "get", 3) || !strncmp(name, "put", 3)) {
+ return true;
+ }
+ }
+ return false;
+}
+
+int MethodData::profile_arguments_flag() {
+ return TypeProfileLevel % 10;
+}
+
+bool MethodData::profile_arguments() {
+ return profile_arguments_flag() > no_type_profile && profile_arguments_flag() <= type_profile_all;
+}
+
+bool MethodData::profile_arguments_jsr292_only() {
+ return profile_arguments_flag() == type_profile_jsr292;
+}
+
+bool MethodData::profile_all_arguments() {
+ return profile_arguments_flag() == type_profile_all;
+}
+
+bool MethodData::profile_arguments_for_invoke(const methodHandle& m, int bci) {
+ if (!profile_arguments()) {
+ return false;
+ }
+
+ if (profile_all_arguments()) {
+ return true;
+ }
+
+ if (profile_unsafe(m, bci)) {
+ return true;
+ }
+
+ assert(profile_arguments_jsr292_only(), "inconsistent");
+ return profile_jsr292(m, bci);
+}
+
+int MethodData::profile_return_flag() {
+ return (TypeProfileLevel % 100) / 10;
+}
+
+bool MethodData::profile_return() {
+ return profile_return_flag() > no_type_profile && profile_return_flag() <= type_profile_all;
+}
+
+bool MethodData::profile_return_jsr292_only() {
+ return profile_return_flag() == type_profile_jsr292;
+}
+
+bool MethodData::profile_all_return() {
+ return profile_return_flag() == type_profile_all;
+}
+
+bool MethodData::profile_return_for_invoke(const methodHandle& m, int bci) {
+ if (!profile_return()) {
+ return false;
+ }
+
+ if (profile_all_return()) {
+ return true;
+ }
+
+ assert(profile_return_jsr292_only(), "inconsistent");
+ return profile_jsr292(m, bci);
+}
+
+int MethodData::profile_parameters_flag() {
+ return TypeProfileLevel / 100;
+}
+
+bool MethodData::profile_parameters() {
+ return profile_parameters_flag() > no_type_profile && profile_parameters_flag() <= type_profile_all;
+}
+
+bool MethodData::profile_parameters_jsr292_only() {
+ return profile_parameters_flag() == type_profile_jsr292;
+}
+
+bool MethodData::profile_all_parameters() {
+ return profile_parameters_flag() == type_profile_all;
+}
+
+bool MethodData::profile_parameters_for_method(const methodHandle& m) {
+ if (!profile_parameters()) {
+ return false;
+ }
+
+ if (profile_all_parameters()) {
+ return true;
+ }
+
+ assert(profile_parameters_jsr292_only(), "inconsistent");
+ return m->is_compiled_lambda_form();
+}
+
+void MethodData::metaspace_pointers_do(MetaspaceClosure* it) {
+ log_trace(cds)("Iter(MethodData): %p", this);
+ it->push(&_method);
+}
+
+void MethodData::clean_extra_data_helper(DataLayout* dp, int shift, bool reset) {
+ if (shift == 0) {
+ return;
+ }
+ if (!reset) {
+ // Move all cells of trap entry at dp left by "shift" cells
+ intptr_t* start = (intptr_t*)dp;
+ intptr_t* end = (intptr_t*)next_extra(dp);
+ for (intptr_t* ptr = start; ptr < end; ptr++) {
+ *(ptr-shift) = *ptr;
+ }
+ } else {
+ // Reset "shift" cells stopping at dp
+ intptr_t* start = ((intptr_t*)dp) - shift;
+ intptr_t* end = (intptr_t*)dp;
+ for (intptr_t* ptr = start; ptr < end; ptr++) {
+ *ptr = 0;
+ }
+ }
+}
+
+class CleanExtraDataClosure : public StackObj {
+public:
+ virtual bool is_live(Method* m) = 0;
+};
+
+// Check for entries that reference an unloaded method
+class CleanExtraDataKlassClosure : public CleanExtraDataClosure {
+private:
+ BoolObjectClosure* _is_alive;
+public:
+ CleanExtraDataKlassClosure(BoolObjectClosure* is_alive) : _is_alive(is_alive) {}
+ bool is_live(Method* m) {
+ return m->method_holder()->is_loader_alive(_is_alive);
+ }
+};
+
+// Check for entries that reference a redefined method
+class CleanExtraDataMethodClosure : public CleanExtraDataClosure {
+public:
+ CleanExtraDataMethodClosure() {}
+ bool is_live(Method* m) { return !m->is_old(); }
+};
+
+
+// Remove SpeculativeTrapData entries that reference an unloaded or
+// redefined method
+void MethodData::clean_extra_data(CleanExtraDataClosure* cl) {
+ DataLayout* dp = extra_data_base();
+ DataLayout* end = args_data_limit();
+
+ int shift = 0;
+ for (; dp < end; dp = next_extra(dp)) {
+ switch(dp->tag()) {
+ case DataLayout::speculative_trap_data_tag: {
+ SpeculativeTrapData* data = new SpeculativeTrapData(dp);
+ Method* m = data->method();
+ assert(m != NULL, "should have a method");
+ if (!cl->is_live(m)) {
+ // "shift" accumulates the number of cells for dead
+ // SpeculativeTrapData entries that have been seen so
+ // far. Following entries must be shifted left by that many
+ // cells to remove the dead SpeculativeTrapData entries.
+ shift += (int)((intptr_t*)next_extra(dp) - (intptr_t*)dp);
+ } else {
+ // Shift this entry left if it follows dead
+ // SpeculativeTrapData entries
+ clean_extra_data_helper(dp, shift);
+ }
+ break;
+ }
+ case DataLayout::bit_data_tag:
+ // Shift this entry left if it follows dead SpeculativeTrapData
+ // entries
+ clean_extra_data_helper(dp, shift);
+ continue;
+ case DataLayout::no_tag:
+ case DataLayout::arg_info_data_tag:
+ // We are at end of the live trap entries. The previous "shift"
+ // cells contain entries that are either dead or were shifted
+ // left. They need to be reset to no_tag
+ clean_extra_data_helper(dp, shift, true);
+ return;
+ default:
+ fatal("unexpected tag %d", dp->tag());
+ }
+ }
+}
+
+// Verify there's no unloaded or redefined method referenced by a
+// SpeculativeTrapData entry
+void MethodData::verify_extra_data_clean(CleanExtraDataClosure* cl) {
+#ifdef ASSERT
+ DataLayout* dp = extra_data_base();
+ DataLayout* end = args_data_limit();
+
+ for (; dp < end; dp = next_extra(dp)) {
+ switch(dp->tag()) {
+ case DataLayout::speculative_trap_data_tag: {
+ SpeculativeTrapData* data = new SpeculativeTrapData(dp);
+ Method* m = data->method();
+ assert(m != NULL && cl->is_live(m), "Method should exist");
+ break;
+ }
+ case DataLayout::bit_data_tag:
+ continue;
+ case DataLayout::no_tag:
+ case DataLayout::arg_info_data_tag:
+ return;
+ default:
+ fatal("unexpected tag %d", dp->tag());
+ }
+ }
+#endif
+}
+
+void MethodData::clean_method_data(BoolObjectClosure* is_alive) {
+ ResourceMark rm;
+ for (ProfileData* data = first_data();
+ is_valid(data);
+ data = next_data(data)) {
+ data->clean_weak_klass_links(is_alive);
+ }
+ ParametersTypeData* parameters = parameters_type_data();
+ if (parameters != NULL) {
+ parameters->clean_weak_klass_links(is_alive);
+ }
+
+ CleanExtraDataKlassClosure cl(is_alive);
+ clean_extra_data(&cl);
+ verify_extra_data_clean(&cl);
+}
+
+void MethodData::clean_weak_method_links() {
+ ResourceMark rm;
+ for (ProfileData* data = first_data();
+ is_valid(data);
+ data = next_data(data)) {
+ data->clean_weak_method_links();
+ }
+
+ CleanExtraDataMethodClosure cl;
+ clean_extra_data(&cl);
+ verify_extra_data_clean(&cl);
+}
+
+#ifdef ASSERT
+void MethodData::verify_clean_weak_method_links() {
+ ResourceMark rm;
+ for (ProfileData* data = first_data();
+ is_valid(data);
+ data = next_data(data)) {
+ data->verify_clean_weak_method_links();
+ }
+
+ CleanExtraDataMethodClosure cl;
+ verify_extra_data_clean(&cl);
+}
+#endif // ASSERT