src/hotspot/share/gc/z/zNMethodTable.cpp
changeset 50525 767cdb97f103
child 52410 9d3abb1b7b1f
--- /dev/null	Thu Jan 01 00:00:00 1970 +0000
+++ b/src/hotspot/share/gc/z/zNMethodTable.cpp	Tue Jun 12 17:40:28 2018 +0200
@@ -0,0 +1,465 @@
+/*
+ * Copyright (c) 2017, 2018, Oracle and/or its affiliates. All rights reserved.
+ * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
+ *
+ * This code is free software; you can redistribute it and/or modify it
+ * under the terms of the GNU General Public License version 2 only, as
+ * published by the Free Software Foundation.
+ *
+ * This code is distributed in the hope that it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
+ * version 2 for more details (a copy is included in the LICENSE file that
+ * accompanied this code).
+ *
+ * You should have received a copy of the GNU General Public License version
+ * 2 along with this work; if not, write to the Free Software Foundation,
+ * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
+ * or visit www.oracle.com if you need additional information or have any
+ * questions.
+ */
+
+#include "precompiled.hpp"
+#include "code/relocInfo.hpp"
+#include "code/nativeInst.hpp"
+#include "code/nmethod.hpp"
+#include "gc/z/zGlobals.hpp"
+#include "gc/z/zHash.inline.hpp"
+#include "gc/z/zNMethodTable.hpp"
+#include "logging/log.hpp"
+#include "memory/allocation.inline.hpp"
+#include "memory/resourceArea.hpp"
+#include "oops/oop.inline.hpp"
+#include "runtime/atomic.hpp"
+#include "utilities/debug.hpp"
+
+class ZNMethodWithImmediateOops {
+private:
+  nmethod* const _nm;
+  const size_t   _nimmediate_oops;
+
+  static size_t header_size();
+
+  ZNMethodWithImmediateOops(nmethod* nm, const GrowableArray<oop*>& immediate_oops);
+
+public:
+  static ZNMethodWithImmediateOops* create(nmethod* nm, const GrowableArray<oop*>& immediate_oops);
+  static void destroy(ZNMethodWithImmediateOops* nmi);
+
+  nmethod* method() const;
+  size_t immediate_oops_count() const;
+  oop** immediate_oops_begin() const;
+  oop** immediate_oops_begin_safe() const;
+  oop** immediate_oops_end() const;
+};
+
+size_t ZNMethodWithImmediateOops::header_size() {
+  const size_t size = sizeof(ZNMethodWithImmediateOops);
+  assert(is_aligned(size, sizeof(oop*)), "Header misaligned");
+  return size;
+}
+
+ZNMethodWithImmediateOops::ZNMethodWithImmediateOops(nmethod* nm, const GrowableArray<oop*>& immediate_oops) :
+    _nm(nm),
+    _nimmediate_oops(immediate_oops.length()) {
+  // Save all immediate oops
+  for (size_t i = 0; i < _nimmediate_oops; i++) {
+    immediate_oops_begin()[i] = immediate_oops.at(i);
+  }
+}
+
+ZNMethodWithImmediateOops* ZNMethodWithImmediateOops::create(nmethod* nm, const GrowableArray<oop*>& immediate_oops) {
+  // Allocate memory for the ZNMethodWithImmediateOops object
+  // plus the immediate oop* array that follows right after.
+  const size_t size = header_size() + (sizeof(oop*) * immediate_oops.length());
+  void* const method_with_immediate_oops = NEW_C_HEAP_ARRAY(uint8_t, size, mtGC);
+  return ::new (method_with_immediate_oops) ZNMethodWithImmediateOops(nm, immediate_oops);
+}
+
+void ZNMethodWithImmediateOops::destroy(ZNMethodWithImmediateOops* nmi) {
+  FREE_C_HEAP_ARRAY(uint8_t, nmi);
+}
+
+nmethod* ZNMethodWithImmediateOops::method() const {
+  return _nm;
+}
+
+size_t ZNMethodWithImmediateOops::immediate_oops_count() const {
+  return _nimmediate_oops;
+}
+
+oop** ZNMethodWithImmediateOops::immediate_oops_begin() const {
+  // The immediate oop* array starts immediately after this object
+  return (oop**)((uintptr_t)this + header_size());
+}
+
+oop** ZNMethodWithImmediateOops::immediate_oops_begin_safe() const {
+  // Non-entrant nmethods have a jump instruction patched into the beginning
+  // of the verified entry point, which could have overwritten an immediate
+  // oop. If so, make sure we skip over that oop.
+  if (_nm->is_not_entrant()) {
+    oop* const first_immediate_oop = *immediate_oops_begin();
+    oop* const safe_begin = (oop*)(_nm->verified_entry_point() + NativeJump::instruction_size);
+    if (first_immediate_oop < safe_begin) {
+      // First immediate oop overwritten, skip it
+      return immediate_oops_begin() + 1;
+    }
+  }
+
+  // First immediate oop not overwritten
+  return immediate_oops_begin();
+}
+
+
+oop** ZNMethodWithImmediateOops::immediate_oops_end() const {
+  return immediate_oops_begin() + immediate_oops_count();
+}
+
+ZNMethodTableEntry* ZNMethodTable::_table = NULL;
+size_t ZNMethodTable::_size = 0;
+size_t ZNMethodTable::_nregistered = 0;
+size_t ZNMethodTable::_nunregistered = 0;
+volatile size_t ZNMethodTable::_claimed = 0;
+
+ZNMethodTableEntry ZNMethodTable::create_entry(nmethod* nm) {
+  GrowableArray<oop*> immediate_oops;
+  bool non_immediate_oops = false;
+
+  // Find all oops relocations
+  RelocIterator iter(nm);
+  while (iter.next()) {
+    if (iter.type() != relocInfo::oop_type) {
+      // Not an oop
+      continue;
+    }
+
+    oop_Relocation* r = iter.oop_reloc();
+
+    if (!r->oop_is_immediate()) {
+      // Non-immediate oop found
+      non_immediate_oops = true;
+      continue;
+    }
+
+    if (r->oop_value() != NULL) {
+      // Non-NULL immediate oop found. NULL oops can safely be
+      // ignored since the method will be re-registered if they
+      // are later patched to be non-NULL.
+      immediate_oops.push(r->oop_addr());
+    }
+  }
+
+  // oops_count() returns the number of oops in the oop table plus one
+  if (immediate_oops.is_empty() && nm->oops_count() == 1) {
+    // No oops found, return empty entry
+    return ZNMethodTableEntry();
+  }
+
+  if (immediate_oops.is_empty()) {
+    // No immediate oops found, return entry without immediate oops
+    return ZNMethodTableEntry(nm, non_immediate_oops);
+  }
+
+  // Return entry with immediate oops
+  return ZNMethodTableEntry(ZNMethodWithImmediateOops::create(nm, immediate_oops), non_immediate_oops);
+}
+
+void ZNMethodTable::destroy_entry(ZNMethodTableEntry entry) {
+  if (entry.immediate_oops()) {
+    ZNMethodWithImmediateOops::destroy(entry.method_with_immediate_oops());
+  }
+}
+
+nmethod* ZNMethodTable::method(ZNMethodTableEntry entry) {
+  return entry.immediate_oops() ? entry.method_with_immediate_oops()->method() : entry.method();
+}
+
+size_t ZNMethodTable::first_index(const nmethod* nm, size_t size) {
+  assert(is_power_of_2(size), "Invalid size");
+  const size_t mask = size - 1;
+  const size_t hash = ZHash::address_to_uint32((uintptr_t)nm);
+  return hash & mask;
+}
+
+size_t ZNMethodTable::next_index(size_t prev_index, size_t size) {
+  assert(is_power_of_2(size), "Invalid size");
+  const size_t mask = size - 1;
+  return (prev_index + 1) & mask;
+}
+
+bool ZNMethodTable::register_entry(ZNMethodTableEntry* table, size_t size, ZNMethodTableEntry entry) {
+  const nmethod* const nm = method(entry);
+  size_t index = first_index(nm, size);
+
+  for (;;) {
+    const ZNMethodTableEntry table_entry = table[index];
+
+    if (!table_entry.registered() && !table_entry.unregistered()) {
+      // Insert new entry
+      table[index] = entry;
+      return true;
+    }
+
+    if (table_entry.registered() && method(table_entry) == nm) {
+      // Replace existing entry
+      destroy_entry(table_entry);
+      table[index] = entry;
+      return false;
+    }
+
+    index = next_index(index, size);
+  }
+}
+
+bool ZNMethodTable::unregister_entry(ZNMethodTableEntry* table, size_t size, const nmethod* nm) {
+  if (size == 0) {
+    // Table is empty
+    return false;
+  }
+
+  size_t index = first_index(nm, size);
+
+  for (;;) {
+    const ZNMethodTableEntry table_entry = table[index];
+
+    if (!table_entry.registered() && !table_entry.unregistered()) {
+      // Entry not found
+      return false;
+    }
+
+    if (table_entry.registered() && method(table_entry) == nm) {
+      // Remove entry
+      destroy_entry(table_entry);
+      table[index] = ZNMethodTableEntry(true /* unregistered */);
+      return true;
+    }
+
+    index = next_index(index, size);
+  }
+}
+
+void ZNMethodTable::rebuild(size_t new_size) {
+  assert(is_power_of_2(new_size), "Invalid size");
+
+  log_debug(gc, nmethod)("Rebuilding NMethod Table: "
+                         SIZE_FORMAT "->" SIZE_FORMAT " entries, "
+                         SIZE_FORMAT "(%.0lf%%->%.0lf%%) registered, "
+                         SIZE_FORMAT "(%.0lf%%->%.0lf%%) unregistered",
+                         _size, new_size,
+                         _nregistered, percent_of(_nregistered, _size), percent_of(_nregistered, new_size),
+                         _nunregistered, percent_of(_nunregistered, _size), 0.0);
+
+  // Allocate new table
+  ZNMethodTableEntry* const new_table = new ZNMethodTableEntry[new_size];
+
+  // Transfer all registered entries
+  for (size_t i = 0; i < _size; i++) {
+    const ZNMethodTableEntry entry = _table[i];
+    if (entry.registered()) {
+      register_entry(new_table, new_size, entry);
+    }
+  }
+
+  // Delete old table
+  delete [] _table;
+
+  // Install new table
+  _table = new_table;
+  _size = new_size;
+  _nunregistered = 0;
+}
+
+void ZNMethodTable::rebuild_if_needed() {
+  // The hash table uses linear probing. To avoid wasting memory while
+  // at the same time maintaining good hash collision behavior we want
+  // to keep the table occupancy between 30% and 70%. The table always
+  // grows/shrinks by doubling/halving its size. Pruning of unregistered
+  // entries is done by rebuilding the table with or without resizing it.
+  const size_t min_size = 1024;
+  const size_t shrink_threshold = _size * 0.30;
+  const size_t prune_threshold = _size * 0.65;
+  const size_t grow_threshold = _size * 0.70;
+
+  if (_size == 0) {
+    // Initialize table
+    rebuild(min_size);
+  } else if (_nregistered < shrink_threshold && _size > min_size) {
+    // Shrink table
+    rebuild(_size / 2);
+  } else if (_nregistered + _nunregistered > grow_threshold) {
+    // Prune or grow table
+    if (_nregistered < prune_threshold) {
+      // Prune table
+      rebuild(_size);
+    } else {
+      // Grow table
+      rebuild(_size * 2);
+    }
+  }
+}
+
+void ZNMethodTable::log_register(const nmethod* nm, ZNMethodTableEntry entry) {
+  LogTarget(Trace, gc, nmethod) log;
+  if (!log.is_enabled()) {
+    return;
+  }
+
+  log.print("Register NMethod: %s.%s (" PTR_FORMAT "), "
+            "Compiler: %s, Oops: %d, ImmediateOops: " SIZE_FORMAT ", NonImmediateOops: %s",
+            nm->method()->method_holder()->external_name(),
+            nm->method()->name()->as_C_string(),
+            p2i(nm),
+            nm->compiler_name(),
+            nm->oops_count() - 1,
+            entry.immediate_oops() ? entry.method_with_immediate_oops()->immediate_oops_count() : 0,
+            BOOL_TO_STR(entry.non_immediate_oops()));
+
+  LogTarget(Trace, gc, nmethod, oops) log_oops;
+  if (!log_oops.is_enabled()) {
+    return;
+  }
+
+  // Print nmethod oops table
+  oop* const begin = nm->oops_begin();
+  oop* const end = nm->oops_end();
+  for (oop* p = begin; p < end; p++) {
+    log_oops.print("           Oop[" SIZE_FORMAT "] " PTR_FORMAT " (%s)",
+                   (p - begin), p2i(*p), (*p)->klass()->external_name());
+  }
+
+  if (entry.immediate_oops()) {
+    // Print nmethod immediate oops
+    const ZNMethodWithImmediateOops* const nmi = entry.method_with_immediate_oops();
+    oop** const begin = nmi->immediate_oops_begin();
+    oop** const end = nmi->immediate_oops_end();
+    for (oop** p = begin; p < end; p++) {
+      log_oops.print("  ImmediateOop[" SIZE_FORMAT "] " PTR_FORMAT " @ " PTR_FORMAT " (%s)",
+                     (p - begin), p2i(**p), p2i(*p), (**p)->klass()->external_name());
+    }
+  }
+}
+
+void ZNMethodTable::log_unregister(const nmethod* nm) {
+  LogTarget(Debug, gc, nmethod) log;
+  if (!log.is_enabled()) {
+    return;
+  }
+
+  log.print("Unregister NMethod: %s.%s (" PTR_FORMAT ")",
+            nm->method()->method_holder()->external_name(),
+            nm->method()->name()->as_C_string(),
+            p2i(nm));
+}
+
+size_t ZNMethodTable::registered_nmethods() {
+  return _nregistered;
+}
+
+size_t ZNMethodTable::unregistered_nmethods() {
+  return _nunregistered;
+}
+
+void ZNMethodTable::register_nmethod(nmethod* nm) {
+  ResourceMark rm;
+
+  // Create entry
+  const ZNMethodTableEntry entry = create_entry(nm);
+
+  log_register(nm, entry);
+
+  if (!entry.registered()) {
+    // Method doesn't have any oops, ignore it
+    return;
+  }
+
+  // Grow/Shrink/Prune table if needed
+  rebuild_if_needed();
+
+  // Insert new entry
+  if (register_entry(_table, _size, entry)) {
+    // New entry registered. When register_entry() instead returns
+    // false the nmethod was already in the table so we do not want
+    // to increase number of registered entries in that case.
+    _nregistered++;
+  }
+}
+
+void ZNMethodTable::unregister_nmethod(nmethod* nm) {
+  ResourceMark rm;
+
+  log_unregister(nm);
+
+  // Remove entry
+  if (unregister_entry(_table, _size, nm)) {
+    // Entry was unregistered. When unregister_entry() instead returns
+    // false the nmethod was not in the table (because it didn't have
+    // any oops) so we do not want to decrease the number of registered
+    // entries in that case.
+    _nregistered--;
+    _nunregistered++;
+  }
+}
+
+void ZNMethodTable::gc_prologue() {
+  _claimed = 0;
+}
+
+void ZNMethodTable::gc_epilogue() {
+  assert(_claimed >= _size, "Failed to claim all table entries");
+}
+
+void ZNMethodTable::entry_oops_do(ZNMethodTableEntry entry, OopClosure* cl) {
+  nmethod* const nm = method(entry);
+  if (!nm->is_alive()) {
+    // No need to visit oops
+    return;
+  }
+
+  // Process oops table
+  oop* const begin = nm->oops_begin();
+  oop* const end = nm->oops_end();
+  for (oop* p = begin; p < end; p++) {
+    if (*p != Universe::non_oop_word()) {
+      cl->do_oop(p);
+    }
+  }
+
+  if (entry.immediate_oops()) {
+    // Process immediate oops
+    const ZNMethodWithImmediateOops* const nmi = entry.method_with_immediate_oops();
+    oop** const begin = nmi->immediate_oops_begin_safe();
+    oop** const end = nmi->immediate_oops_end();
+    for (oop** p = begin; p < end; p++) {
+      cl->do_oop(*p);
+    }
+  }
+
+  if (entry.non_immediate_oops()) {
+    // Process non-immediate oops
+    nm->fix_oop_relocations();
+  }
+}
+
+void ZNMethodTable::oops_do(OopClosure* cl) {
+  for (;;) {
+    // Claim table partition. Each partition is currently sized to span
+    // two cache lines. This number is just a guess, but seems to work well.
+    const size_t partition_size = (ZCacheLineSize * 2) / sizeof(ZNMethodTableEntry);
+    const size_t partition_start = MIN2(Atomic::add(partition_size, &_claimed) - partition_size, _size);
+    const size_t partition_end = MIN2(partition_start + partition_size, _size);
+    if (partition_start == partition_end) {
+      // End of table
+      break;
+    }
+
+    // Process table partition
+    for (size_t i = partition_start; i < partition_end; i++) {
+      const ZNMethodTableEntry entry = _table[i];
+      if (entry.registered()) {
+        entry_oops_do(entry, cl);
+      }
+    }
+  }
+}