src/hotspot/share/gc/parallel/gcTaskManager.cpp
changeset 47216 71c04702a3d5
parent 46701 f559541c0daa
child 50429 83aec1d357d4
--- /dev/null	Thu Jan 01 00:00:00 1970 +0000
+++ b/src/hotspot/share/gc/parallel/gcTaskManager.cpp	Tue Sep 12 19:03:39 2017 +0200
@@ -0,0 +1,1073 @@
+/*
+ * Copyright (c) 2002, 2016, Oracle and/or its affiliates. All rights reserved.
+ * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
+ *
+ * This code is free software; you can redistribute it and/or modify it
+ * under the terms of the GNU General Public License version 2 only, as
+ * published by the Free Software Foundation.
+ *
+ * This code is distributed in the hope that it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
+ * version 2 for more details (a copy is included in the LICENSE file that
+ * accompanied this code).
+ *
+ * You should have received a copy of the GNU General Public License version
+ * 2 along with this work; if not, write to the Free Software Foundation,
+ * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
+ * or visit www.oracle.com if you need additional information or have any
+ * questions.
+ *
+ */
+
+#include "precompiled.hpp"
+#include "gc/parallel/gcTaskManager.hpp"
+#include "gc/parallel/gcTaskThread.hpp"
+#include "gc/shared/gcId.hpp"
+#include "gc/shared/workerManager.hpp"
+#include "logging/log.hpp"
+#include "logging/logStream.hpp"
+#include "memory/allocation.hpp"
+#include "memory/allocation.inline.hpp"
+#include "memory/resourceArea.hpp"
+#include "runtime/mutex.hpp"
+#include "runtime/mutexLocker.hpp"
+#include "runtime/orderAccess.inline.hpp"
+#include "runtime/os.hpp"
+
+//
+// GCTask
+//
+
+const char* GCTask::Kind::to_string(kind value) {
+  const char* result = "unknown GCTask kind";
+  switch (value) {
+  default:
+    result = "unknown GCTask kind";
+    break;
+  case unknown_task:
+    result = "unknown task";
+    break;
+  case ordinary_task:
+    result = "ordinary task";
+    break;
+  case wait_for_barrier_task:
+    result = "wait for barrier task";
+    break;
+  case noop_task:
+    result = "noop task";
+    break;
+  case idle_task:
+    result = "idle task";
+    break;
+  }
+  return result;
+};
+
+GCTask::GCTask() {
+  initialize(Kind::ordinary_task, GCId::current());
+}
+
+GCTask::GCTask(Kind::kind kind) {
+  initialize(kind, GCId::current());
+}
+
+GCTask::GCTask(Kind::kind kind, uint gc_id) {
+  initialize(kind, gc_id);
+}
+
+void GCTask::initialize(Kind::kind kind, uint gc_id) {
+  _kind = kind;
+  _affinity = GCTaskManager::sentinel_worker();
+  _older = NULL;
+  _newer = NULL;
+  _gc_id = gc_id;
+}
+
+void GCTask::destruct() {
+  assert(older() == NULL, "shouldn't have an older task");
+  assert(newer() == NULL, "shouldn't have a newer task");
+  // Nothing to do.
+}
+
+NOT_PRODUCT(
+void GCTask::print(const char* message) const {
+  tty->print(INTPTR_FORMAT " <- " INTPTR_FORMAT "(%u) -> " INTPTR_FORMAT,
+             p2i(newer()), p2i(this), affinity(), p2i(older()));
+}
+)
+
+//
+// GCTaskQueue
+//
+
+GCTaskQueue* GCTaskQueue::create() {
+  GCTaskQueue* result = new GCTaskQueue(false);
+  if (TraceGCTaskQueue) {
+    tty->print_cr("GCTaskQueue::create()"
+                  " returns " INTPTR_FORMAT, p2i(result));
+  }
+  return result;
+}
+
+GCTaskQueue* GCTaskQueue::create_on_c_heap() {
+  GCTaskQueue* result = new(ResourceObj::C_HEAP, mtGC) GCTaskQueue(true);
+  if (TraceGCTaskQueue) {
+    tty->print_cr("GCTaskQueue::create_on_c_heap()"
+                  " returns " INTPTR_FORMAT,
+                  p2i(result));
+  }
+  return result;
+}
+
+GCTaskQueue::GCTaskQueue(bool on_c_heap) :
+  _is_c_heap_obj(on_c_heap) {
+  initialize();
+  if (TraceGCTaskQueue) {
+    tty->print_cr("[" INTPTR_FORMAT "]"
+                  " GCTaskQueue::GCTaskQueue() constructor",
+                  p2i(this));
+  }
+}
+
+void GCTaskQueue::destruct() {
+  // Nothing to do.
+}
+
+void GCTaskQueue::destroy(GCTaskQueue* that) {
+  if (TraceGCTaskQueue) {
+    tty->print_cr("[" INTPTR_FORMAT "]"
+                  " GCTaskQueue::destroy()"
+                  "  is_c_heap_obj:  %s",
+                  p2i(that),
+                  that->is_c_heap_obj() ? "true" : "false");
+  }
+  // That instance may have been allocated as a CHeapObj,
+  // in which case we have to free it explicitly.
+  if (that != NULL) {
+    that->destruct();
+    assert(that->is_empty(), "should be empty");
+    if (that->is_c_heap_obj()) {
+      FreeHeap(that);
+    }
+  }
+}
+
+void GCTaskQueue::initialize() {
+  set_insert_end(NULL);
+  set_remove_end(NULL);
+  set_length(0);
+}
+
+// Enqueue one task.
+void GCTaskQueue::enqueue(GCTask* task) {
+  if (TraceGCTaskQueue) {
+    tty->print_cr("[" INTPTR_FORMAT "]"
+                  " GCTaskQueue::enqueue(task: "
+                  INTPTR_FORMAT ")",
+                  p2i(this), p2i(task));
+    print("before:");
+  }
+  assert(task != NULL, "shouldn't have null task");
+  assert(task->older() == NULL, "shouldn't be on queue");
+  assert(task->newer() == NULL, "shouldn't be on queue");
+  task->set_newer(NULL);
+  task->set_older(insert_end());
+  if (is_empty()) {
+    set_remove_end(task);
+  } else {
+    insert_end()->set_newer(task);
+  }
+  set_insert_end(task);
+  increment_length();
+  verify_length();
+  if (TraceGCTaskQueue) {
+    print("after:");
+  }
+}
+
+// Enqueue a whole list of tasks.  Empties the argument list.
+void GCTaskQueue::enqueue(GCTaskQueue* list) {
+  if (TraceGCTaskQueue) {
+    tty->print_cr("[" INTPTR_FORMAT "]"
+                  " GCTaskQueue::enqueue(list: "
+                  INTPTR_FORMAT ")",
+                  p2i(this), p2i(list));
+    print("before:");
+    list->print("list:");
+  }
+  if (list->is_empty()) {
+    // Enqueueing the empty list: nothing to do.
+    return;
+  }
+  uint list_length = list->length();
+  if (is_empty()) {
+    // Enqueueing to empty list: just acquire elements.
+    set_insert_end(list->insert_end());
+    set_remove_end(list->remove_end());
+    set_length(list_length);
+  } else {
+    // Prepend argument list to our queue.
+    list->remove_end()->set_older(insert_end());
+    insert_end()->set_newer(list->remove_end());
+    set_insert_end(list->insert_end());
+    set_length(length() + list_length);
+    // empty the argument list.
+  }
+  list->initialize();
+  if (TraceGCTaskQueue) {
+    print("after:");
+    list->print("list:");
+  }
+  verify_length();
+}
+
+// Dequeue one task.
+GCTask* GCTaskQueue::dequeue() {
+  if (TraceGCTaskQueue) {
+    tty->print_cr("[" INTPTR_FORMAT "]"
+                  " GCTaskQueue::dequeue()", p2i(this));
+    print("before:");
+  }
+  assert(!is_empty(), "shouldn't dequeue from empty list");
+  GCTask* result = remove();
+  assert(result != NULL, "shouldn't have NULL task");
+  if (TraceGCTaskQueue) {
+    tty->print_cr("    return: " INTPTR_FORMAT, p2i(result));
+    print("after:");
+  }
+  return result;
+}
+
+// Dequeue one task, preferring one with affinity.
+GCTask* GCTaskQueue::dequeue(uint affinity) {
+  if (TraceGCTaskQueue) {
+    tty->print_cr("[" INTPTR_FORMAT "]"
+                  " GCTaskQueue::dequeue(%u)", p2i(this), affinity);
+    print("before:");
+  }
+  assert(!is_empty(), "shouldn't dequeue from empty list");
+  // Look down to the next barrier for a task with this affinity.
+  GCTask* result = NULL;
+  for (GCTask* element = remove_end();
+       element != NULL;
+       element = element->newer()) {
+    if (element->is_barrier_task()) {
+      // Don't consider barrier tasks, nor past them.
+      result = NULL;
+      break;
+    }
+    if (element->affinity() == affinity) {
+      result = remove(element);
+      break;
+    }
+  }
+  // If we didn't find anything with affinity, just take the next task.
+  if (result == NULL) {
+    result = remove();
+  }
+  if (TraceGCTaskQueue) {
+    tty->print_cr("    return: " INTPTR_FORMAT, p2i(result));
+    print("after:");
+  }
+  return result;
+}
+
+GCTask* GCTaskQueue::remove() {
+  // Dequeue from remove end.
+  GCTask* result = remove_end();
+  assert(result != NULL, "shouldn't have null task");
+  assert(result->older() == NULL, "not the remove_end");
+  set_remove_end(result->newer());
+  if (remove_end() == NULL) {
+    assert(insert_end() == result, "not a singleton");
+    set_insert_end(NULL);
+  } else {
+    remove_end()->set_older(NULL);
+  }
+  result->set_newer(NULL);
+  decrement_length();
+  assert(result->newer() == NULL, "shouldn't be on queue");
+  assert(result->older() == NULL, "shouldn't be on queue");
+  verify_length();
+  return result;
+}
+
+GCTask* GCTaskQueue::remove(GCTask* task) {
+  // This is slightly more work, and has slightly fewer asserts
+  // than removing from the remove end.
+  assert(task != NULL, "shouldn't have null task");
+  GCTask* result = task;
+  if (result->newer() != NULL) {
+    result->newer()->set_older(result->older());
+  } else {
+    assert(insert_end() == result, "not youngest");
+    set_insert_end(result->older());
+  }
+  if (result->older() != NULL) {
+    result->older()->set_newer(result->newer());
+  } else {
+    assert(remove_end() == result, "not oldest");
+    set_remove_end(result->newer());
+  }
+  result->set_newer(NULL);
+  result->set_older(NULL);
+  decrement_length();
+  verify_length();
+  return result;
+}
+
+NOT_PRODUCT(
+// Count the elements in the queue and verify the length against
+// that count.
+void GCTaskQueue::verify_length() const {
+  uint count = 0;
+  for (GCTask* element = insert_end();
+       element != NULL;
+       element = element->older()) {
+
+    count++;
+  }
+  assert(count == length(), "Length does not match queue");
+}
+
+void GCTaskQueue::print(const char* message) const {
+  tty->print_cr("[" INTPTR_FORMAT "] GCTaskQueue:"
+                "  insert_end: " INTPTR_FORMAT
+                "  remove_end: " INTPTR_FORMAT
+                "  length:       %d"
+                "  %s",
+                p2i(this), p2i(insert_end()), p2i(remove_end()), length(), message);
+  uint count = 0;
+  for (GCTask* element = insert_end();
+       element != NULL;
+       element = element->older()) {
+    element->print("    ");
+    count++;
+    tty->cr();
+  }
+  tty->print("Total tasks: %d", count);
+}
+)
+
+//
+// SynchronizedGCTaskQueue
+//
+
+SynchronizedGCTaskQueue::SynchronizedGCTaskQueue(GCTaskQueue* queue_arg,
+                                                 Monitor *       lock_arg) :
+  _unsynchronized_queue(queue_arg),
+  _lock(lock_arg) {
+  assert(unsynchronized_queue() != NULL, "null queue");
+  assert(lock() != NULL, "null lock");
+}
+
+SynchronizedGCTaskQueue::~SynchronizedGCTaskQueue() {
+  // Nothing to do.
+}
+
+//
+// GCTaskManager
+//
+GCTaskManager::GCTaskManager(uint workers) :
+  _workers(workers),
+  _active_workers(0),
+  _idle_workers(0),
+  _created_workers(0) {
+  initialize();
+}
+
+GCTaskThread* GCTaskManager::install_worker(uint t) {
+  GCTaskThread* new_worker = GCTaskThread::create(this, t, _processor_assignment[t]);
+  set_thread(t, new_worker);
+  return new_worker;
+}
+
+void GCTaskManager::add_workers(bool initializing) {
+  os::ThreadType worker_type = os::pgc_thread;
+  uint previous_created_workers = _created_workers;
+
+  _created_workers = WorkerManager::add_workers(this,
+                                                _active_workers,
+                                                _workers,
+                                                _created_workers,
+                                                worker_type,
+                                                initializing);
+  _active_workers = MIN2(_created_workers, _active_workers);
+
+  WorkerManager::log_worker_creation(this, previous_created_workers, _active_workers, _created_workers, initializing);
+}
+
+const char* GCTaskManager::group_name() {
+  return "ParGC Thread";
+}
+
+void GCTaskManager::initialize() {
+  if (TraceGCTaskManager) {
+    tty->print_cr("GCTaskManager::initialize: workers: %u", workers());
+  }
+  assert(workers() != 0, "no workers");
+  _monitor = new Monitor(Mutex::barrier,                // rank
+                         "GCTaskManager monitor",       // name
+                         Mutex::_allow_vm_block_flag,   // allow_vm_block
+                         Monitor::_safepoint_check_never);
+  // The queue for the GCTaskManager must be a CHeapObj.
+  GCTaskQueue* unsynchronized_queue = GCTaskQueue::create_on_c_heap();
+  _queue = SynchronizedGCTaskQueue::create(unsynchronized_queue, lock());
+  _noop_task = NoopGCTask::create_on_c_heap();
+  _resource_flag = NEW_C_HEAP_ARRAY(bool, workers(), mtGC);
+  {
+    // Set up worker threads.
+    //     Distribute the workers among the available processors,
+    //     unless we were told not to, or if the os doesn't want to.
+    _processor_assignment = NEW_C_HEAP_ARRAY(uint, workers(), mtGC);
+    if (!BindGCTaskThreadsToCPUs ||
+        !os::distribute_processes(workers(), _processor_assignment)) {
+      for (uint a = 0; a < workers(); a += 1) {
+        _processor_assignment[a] = sentinel_worker();
+      }
+    }
+
+    _thread = NEW_C_HEAP_ARRAY(GCTaskThread*, workers(), mtGC);
+    _active_workers = ParallelGCThreads;
+    if (UseDynamicNumberOfGCThreads && !FLAG_IS_CMDLINE(ParallelGCThreads)) {
+      _active_workers = 1U;
+    }
+
+    Log(gc, task, thread) log;
+    if (log.is_trace()) {
+      LogStream ls(log.trace());
+      ls.print("GCTaskManager::initialize: distribution:");
+      for (uint t = 0; t < workers(); t += 1) {
+        ls.print("  %u", _processor_assignment[t]);
+      }
+      ls.cr();
+    }
+  }
+  reset_busy_workers();
+  set_unblocked();
+  for (uint w = 0; w < workers(); w += 1) {
+    set_resource_flag(w, false);
+  }
+  reset_delivered_tasks();
+  reset_completed_tasks();
+  reset_barriers();
+  reset_emptied_queue();
+
+  add_workers(true);
+}
+
+GCTaskManager::~GCTaskManager() {
+  assert(busy_workers() == 0, "still have busy workers");
+  assert(queue()->is_empty(), "still have queued work");
+  NoopGCTask::destroy(_noop_task);
+  _noop_task = NULL;
+  if (_thread != NULL) {
+    for (uint i = 0; i < created_workers(); i += 1) {
+      GCTaskThread::destroy(thread(i));
+      set_thread(i, NULL);
+    }
+    FREE_C_HEAP_ARRAY(GCTaskThread*, _thread);
+    _thread = NULL;
+  }
+  if (_processor_assignment != NULL) {
+    FREE_C_HEAP_ARRAY(uint, _processor_assignment);
+    _processor_assignment = NULL;
+  }
+  if (_resource_flag != NULL) {
+    FREE_C_HEAP_ARRAY(bool, _resource_flag);
+    _resource_flag = NULL;
+  }
+  if (queue() != NULL) {
+    GCTaskQueue* unsynchronized_queue = queue()->unsynchronized_queue();
+    GCTaskQueue::destroy(unsynchronized_queue);
+    SynchronizedGCTaskQueue::destroy(queue());
+    _queue = NULL;
+  }
+  if (monitor() != NULL) {
+    delete monitor();
+    _monitor = NULL;
+  }
+}
+
+void GCTaskManager::set_active_gang() {
+  _active_workers =
+    AdaptiveSizePolicy::calc_active_workers(workers(),
+                                 active_workers(),
+                                 Threads::number_of_non_daemon_threads());
+
+  assert(!all_workers_active() || active_workers() == ParallelGCThreads,
+         "all_workers_active() is  incorrect: "
+         "active %d  ParallelGCThreads %u", active_workers(),
+         ParallelGCThreads);
+  _active_workers = MIN2(_active_workers, _workers);
+  // "add_workers" does not guarantee any additional workers
+  add_workers(false);
+  log_trace(gc, task)("GCTaskManager::set_active_gang(): "
+                      "all_workers_active()  %d  workers %d  "
+                      "active  %d  ParallelGCThreads %u",
+                      all_workers_active(), workers(),  active_workers(),
+                      ParallelGCThreads);
+}
+
+// Create IdleGCTasks for inactive workers.
+// Creates tasks in a ResourceArea and assumes
+// an appropriate ResourceMark.
+void GCTaskManager::task_idle_workers() {
+  {
+    int more_inactive_workers = 0;
+    {
+      // Stop any idle tasks from exiting their IdleGCTask's
+      // and get the count for additional IdleGCTask's under
+      // the GCTaskManager's monitor so that the "more_inactive_workers"
+      // count is correct.
+      MutexLockerEx ml(monitor(), Mutex::_no_safepoint_check_flag);
+      _wait_helper.set_should_wait(true);
+      // active_workers are a number being requested.  idle_workers
+      // are the number currently idle.  If all the workers are being
+      // requested to be active but some are already idle, reduce
+      // the number of active_workers to be consistent with the
+      // number of idle_workers.  The idle_workers are stuck in
+      // idle tasks and will no longer be release (since a new GC
+      // is starting).  Try later to release enough idle_workers
+      // to allow the desired number of active_workers.
+      more_inactive_workers =
+        created_workers() - active_workers() - idle_workers();
+      if (more_inactive_workers < 0) {
+        int reduced_active_workers = active_workers() + more_inactive_workers;
+        update_active_workers(reduced_active_workers);
+        more_inactive_workers = 0;
+      }
+      log_trace(gc, task)("JT: %d  workers %d  active  %d  idle %d  more %d",
+                          Threads::number_of_non_daemon_threads(),
+                          created_workers(),
+                          active_workers(),
+                          idle_workers(),
+                          more_inactive_workers);
+    }
+    GCTaskQueue* q = GCTaskQueue::create();
+    for(uint i = 0; i < (uint) more_inactive_workers; i++) {
+      q->enqueue(IdleGCTask::create_on_c_heap());
+      increment_idle_workers();
+    }
+    assert(created_workers() == active_workers() + idle_workers(),
+      "total workers should equal active + inactive");
+    add_list(q);
+    // GCTaskQueue* q was created in a ResourceArea so a
+    // destroy() call is not needed.
+  }
+}
+
+void  GCTaskManager::release_idle_workers() {
+  {
+    MutexLockerEx ml(monitor(),
+      Mutex::_no_safepoint_check_flag);
+    _wait_helper.set_should_wait(false);
+    monitor()->notify_all();
+  // Release monitor
+  }
+}
+
+void GCTaskManager::print_task_time_stamps() {
+  if (!log_is_enabled(Debug, gc, task, time)) {
+    return;
+  }
+  uint num_thr = created_workers();
+  for(uint i=0; i < num_thr; i++) {
+    GCTaskThread* t = thread(i);
+    t->print_task_time_stamps();
+  }
+}
+
+void GCTaskManager::print_threads_on(outputStream* st) {
+  uint num_thr = created_workers();
+  for (uint i = 0; i < num_thr; i++) {
+    thread(i)->print_on(st);
+    st->cr();
+  }
+}
+
+void GCTaskManager::threads_do(ThreadClosure* tc) {
+  assert(tc != NULL, "Null ThreadClosure");
+  uint num_thr = created_workers();
+  for (uint i = 0; i < num_thr; i++) {
+    tc->do_thread(thread(i));
+  }
+}
+
+GCTaskThread* GCTaskManager::thread(uint which) {
+  assert(which < created_workers(), "index out of bounds");
+  assert(_thread[which] != NULL, "shouldn't have null thread");
+  return _thread[which];
+}
+
+void GCTaskManager::set_thread(uint which, GCTaskThread* value) {
+  // "_created_workers" may not have been updated yet so use workers()
+  assert(which < workers(), "index out of bounds");
+  assert(value != NULL, "shouldn't have null thread");
+  _thread[which] = value;
+}
+
+void GCTaskManager::add_task(GCTask* task) {
+  assert(task != NULL, "shouldn't have null task");
+  MutexLockerEx ml(monitor(), Mutex::_no_safepoint_check_flag);
+  if (TraceGCTaskManager) {
+    tty->print_cr("GCTaskManager::add_task(" INTPTR_FORMAT " [%s])",
+                  p2i(task), GCTask::Kind::to_string(task->kind()));
+  }
+  queue()->enqueue(task);
+  // Notify with the lock held to avoid missed notifies.
+  if (TraceGCTaskManager) {
+    tty->print_cr("    GCTaskManager::add_task (%s)->notify_all",
+                  monitor()->name());
+  }
+  (void) monitor()->notify_all();
+  // Release monitor().
+}
+
+void GCTaskManager::add_list(GCTaskQueue* list) {
+  assert(list != NULL, "shouldn't have null task");
+  MutexLockerEx ml(monitor(), Mutex::_no_safepoint_check_flag);
+  if (TraceGCTaskManager) {
+    tty->print_cr("GCTaskManager::add_list(%u)", list->length());
+  }
+  queue()->enqueue(list);
+  // Notify with the lock held to avoid missed notifies.
+  if (TraceGCTaskManager) {
+    tty->print_cr("    GCTaskManager::add_list (%s)->notify_all",
+                  monitor()->name());
+  }
+  (void) monitor()->notify_all();
+  // Release monitor().
+}
+
+// GC workers wait in get_task() for new work to be added
+// to the GCTaskManager's queue.  When new work is added,
+// a notify is sent to the waiting GC workers which then
+// compete to get tasks.  If a GC worker wakes up and there
+// is no work on the queue, it is given a noop_task to execute
+// and then loops to find more work.
+
+GCTask* GCTaskManager::get_task(uint which) {
+  GCTask* result = NULL;
+  // Grab the queue lock.
+  MutexLockerEx ml(monitor(), Mutex::_no_safepoint_check_flag);
+  // Wait while the queue is block or
+  // there is nothing to do, except maybe release resources.
+  while (is_blocked() ||
+         (queue()->is_empty() && !should_release_resources(which))) {
+    if (TraceGCTaskManager) {
+      tty->print_cr("GCTaskManager::get_task(%u)"
+                    "  blocked: %s"
+                    "  empty: %s"
+                    "  release: %s",
+                    which,
+                    is_blocked() ? "true" : "false",
+                    queue()->is_empty() ? "true" : "false",
+                    should_release_resources(which) ? "true" : "false");
+      tty->print_cr("    => (%s)->wait()",
+                    monitor()->name());
+    }
+    monitor()->wait(Mutex::_no_safepoint_check_flag, 0);
+  }
+  // We've reacquired the queue lock here.
+  // Figure out which condition caused us to exit the loop above.
+  if (!queue()->is_empty()) {
+    if (UseGCTaskAffinity) {
+      result = queue()->dequeue(which);
+    } else {
+      result = queue()->dequeue();
+    }
+    if (result->is_barrier_task()) {
+      assert(which != sentinel_worker(),
+             "blocker shouldn't be bogus");
+      set_blocking_worker(which);
+    }
+  } else {
+    // The queue is empty, but we were woken up.
+    // Just hand back a Noop task,
+    // in case someone wanted us to release resources, or whatever.
+    result = noop_task();
+  }
+  assert(result != NULL, "shouldn't have null task");
+  if (TraceGCTaskManager) {
+    tty->print_cr("GCTaskManager::get_task(%u) => " INTPTR_FORMAT " [%s]",
+                  which, p2i(result), GCTask::Kind::to_string(result->kind()));
+    tty->print_cr("     %s", result->name());
+  }
+  if (!result->is_idle_task()) {
+    increment_busy_workers();
+    increment_delivered_tasks();
+  }
+  return result;
+  // Release monitor().
+}
+
+void GCTaskManager::note_completion(uint which) {
+  MutexLockerEx ml(monitor(), Mutex::_no_safepoint_check_flag);
+  if (TraceGCTaskManager) {
+    tty->print_cr("GCTaskManager::note_completion(%u)", which);
+  }
+  // If we are blocked, check if the completing thread is the blocker.
+  if (blocking_worker() == which) {
+    assert(blocking_worker() != sentinel_worker(),
+           "blocker shouldn't be bogus");
+    increment_barriers();
+    set_unblocked();
+  }
+  increment_completed_tasks();
+  uint active = decrement_busy_workers();
+  if ((active == 0) && (queue()->is_empty())) {
+    increment_emptied_queue();
+    if (TraceGCTaskManager) {
+      tty->print_cr("    GCTaskManager::note_completion(%u) done", which);
+    }
+  }
+  if (TraceGCTaskManager) {
+    tty->print_cr("    GCTaskManager::note_completion(%u) (%s)->notify_all",
+                  which, monitor()->name());
+    tty->print_cr("  "
+                  "  blocked: %s"
+                  "  empty: %s"
+                  "  release: %s",
+                  is_blocked() ? "true" : "false",
+                  queue()->is_empty() ? "true" : "false",
+                  should_release_resources(which) ? "true" : "false");
+    tty->print_cr("  "
+                  "  delivered: %u"
+                  "  completed: %u"
+                  "  barriers: %u"
+                  "  emptied: %u",
+                  delivered_tasks(),
+                  completed_tasks(),
+                  barriers(),
+                  emptied_queue());
+  }
+  // Tell everyone that a task has completed.
+  (void) monitor()->notify_all();
+  // Release monitor().
+}
+
+uint GCTaskManager::increment_busy_workers() {
+  assert(queue()->own_lock(), "don't own the lock");
+  _busy_workers += 1;
+  return _busy_workers;
+}
+
+uint GCTaskManager::decrement_busy_workers() {
+  assert(queue()->own_lock(), "don't own the lock");
+  assert(_busy_workers > 0, "About to make a mistake");
+  _busy_workers -= 1;
+  return _busy_workers;
+}
+
+void GCTaskManager::release_all_resources() {
+  // If you want this to be done atomically, do it in a WaitForBarrierGCTask.
+  for (uint i = 0; i < created_workers(); i += 1) {
+    set_resource_flag(i, true);
+  }
+}
+
+bool GCTaskManager::should_release_resources(uint which) {
+  // This can be done without a lock because each thread reads one element.
+  return resource_flag(which);
+}
+
+void GCTaskManager::note_release(uint which) {
+  // This can be done without a lock because each thread writes one element.
+  set_resource_flag(which, false);
+}
+
+// "list" contains tasks that are ready to execute.  Those
+// tasks are added to the GCTaskManager's queue of tasks and
+// then the GC workers are notified that there is new work to
+// do.
+//
+// Typically different types of tasks can be added to the "list".
+// For example in PSScavenge OldToYoungRootsTask, SerialOldToYoungRootsTask,
+// ScavengeRootsTask, and StealTask tasks are all added to the list
+// and then the GC workers are notified of new work.  The tasks are
+// handed out in the order in which they are added to the list
+// (although execution is not necessarily in that order).  As long
+// as any tasks are running the GCTaskManager will wait for execution
+// to complete.  GC workers that execute a stealing task remain in
+// the stealing task until all stealing tasks have completed.  The load
+// balancing afforded by the stealing tasks work best if the stealing
+// tasks are added last to the list.
+
+void GCTaskManager::execute_and_wait(GCTaskQueue* list) {
+  WaitForBarrierGCTask* fin = WaitForBarrierGCTask::create();
+  list->enqueue(fin);
+  // The barrier task will be read by one of the GC
+  // workers once it is added to the list of tasks.
+  // Be sure that is globally visible before the
+  // GC worker reads it (which is after the task is added
+  // to the list of tasks below).
+  OrderAccess::storestore();
+  add_list(list);
+  fin->wait_for(true /* reset */);
+  // We have to release the barrier tasks!
+  WaitForBarrierGCTask::destroy(fin);
+}
+
+bool GCTaskManager::resource_flag(uint which) {
+  assert(which < workers(), "index out of bounds");
+  return _resource_flag[which];
+}
+
+void GCTaskManager::set_resource_flag(uint which, bool value) {
+  assert(which < workers(), "index out of bounds");
+  _resource_flag[which] = value;
+}
+
+//
+// NoopGCTask
+//
+
+NoopGCTask* NoopGCTask::create_on_c_heap() {
+  NoopGCTask* result = new(ResourceObj::C_HEAP, mtGC) NoopGCTask();
+  return result;
+}
+
+void NoopGCTask::destroy(NoopGCTask* that) {
+  if (that != NULL) {
+    that->destruct();
+    FreeHeap(that);
+  }
+}
+
+// This task should never be performing GC work that require
+// a valid GC id.
+NoopGCTask::NoopGCTask() : GCTask(GCTask::Kind::noop_task, GCId::undefined()) { }
+
+void NoopGCTask::destruct() {
+  // This has to know it's superclass structure, just like the constructor.
+  this->GCTask::destruct();
+  // Nothing else to do.
+}
+
+//
+// IdleGCTask
+//
+
+IdleGCTask* IdleGCTask::create() {
+  IdleGCTask* result = new IdleGCTask(false);
+  assert(UseDynamicNumberOfGCThreads,
+    "Should only be used with dynamic GC thread");
+  return result;
+}
+
+IdleGCTask* IdleGCTask::create_on_c_heap() {
+  IdleGCTask* result = new(ResourceObj::C_HEAP, mtGC) IdleGCTask(true);
+  assert(UseDynamicNumberOfGCThreads,
+    "Should only be used with dynamic GC thread");
+  return result;
+}
+
+void IdleGCTask::do_it(GCTaskManager* manager, uint which) {
+  WaitHelper* wait_helper = manager->wait_helper();
+  log_trace(gc, task)("[" INTPTR_FORMAT "] IdleGCTask:::do_it() should_wait: %s",
+      p2i(this), wait_helper->should_wait() ? "true" : "false");
+
+  MutexLockerEx ml(manager->monitor(), Mutex::_no_safepoint_check_flag);
+  log_trace(gc, task)("--- idle %d", which);
+  // Increment has to be done when the idle tasks are created.
+  // manager->increment_idle_workers();
+  manager->monitor()->notify_all();
+  while (wait_helper->should_wait()) {
+    log_trace(gc, task)("[" INTPTR_FORMAT "] IdleGCTask::do_it()  [" INTPTR_FORMAT "] (%s)->wait()",
+      p2i(this), p2i(manager->monitor()), manager->monitor()->name());
+    manager->monitor()->wait(Mutex::_no_safepoint_check_flag, 0);
+  }
+  manager->decrement_idle_workers();
+
+  log_trace(gc, task)("--- release %d", which);
+  log_trace(gc, task)("[" INTPTR_FORMAT "] IdleGCTask::do_it() returns should_wait: %s",
+    p2i(this), wait_helper->should_wait() ? "true" : "false");
+  // Release monitor().
+}
+
+void IdleGCTask::destroy(IdleGCTask* that) {
+  if (that != NULL) {
+    that->destruct();
+    if (that->is_c_heap_obj()) {
+      FreeHeap(that);
+    }
+  }
+}
+
+void IdleGCTask::destruct() {
+  // This has to know it's superclass structure, just like the constructor.
+  this->GCTask::destruct();
+  // Nothing else to do.
+}
+
+//
+// WaitForBarrierGCTask
+//
+WaitForBarrierGCTask* WaitForBarrierGCTask::create() {
+  WaitForBarrierGCTask* result = new WaitForBarrierGCTask();
+  return result;
+}
+
+WaitForBarrierGCTask::WaitForBarrierGCTask() : GCTask(GCTask::Kind::wait_for_barrier_task) { }
+
+void WaitForBarrierGCTask::destroy(WaitForBarrierGCTask* that) {
+  if (that != NULL) {
+    if (TraceGCTaskManager) {
+      tty->print_cr("[" INTPTR_FORMAT "] WaitForBarrierGCTask::destroy()", p2i(that));
+    }
+    that->destruct();
+  }
+}
+
+void WaitForBarrierGCTask::destruct() {
+  if (TraceGCTaskManager) {
+    tty->print_cr("[" INTPTR_FORMAT "] WaitForBarrierGCTask::destruct()", p2i(this));
+  }
+  this->GCTask::destruct();
+  // Clean up that should be in the destructor,
+  // except that ResourceMarks don't call destructors.
+  _wait_helper.release_monitor();
+}
+
+void WaitForBarrierGCTask::do_it_internal(GCTaskManager* manager, uint which) {
+  // Wait for this to be the only busy worker.
+  assert(manager->monitor()->owned_by_self(), "don't own the lock");
+  assert(manager->is_blocked(), "manager isn't blocked");
+  while (manager->busy_workers() > 1) {
+    if (TraceGCTaskManager) {
+      tty->print_cr("WaitForBarrierGCTask::do_it(%u) waiting on %u workers",
+                    which, manager->busy_workers());
+    }
+    manager->monitor()->wait(Mutex::_no_safepoint_check_flag, 0);
+  }
+}
+
+void WaitForBarrierGCTask::do_it(GCTaskManager* manager, uint which) {
+  if (TraceGCTaskManager) {
+    tty->print_cr("[" INTPTR_FORMAT "]"
+                  " WaitForBarrierGCTask::do_it() waiting for idle",
+                  p2i(this));
+  }
+  {
+    // First, wait for the barrier to arrive.
+    MutexLockerEx ml(manager->lock(), Mutex::_no_safepoint_check_flag);
+    do_it_internal(manager, which);
+    // Release manager->lock().
+  }
+  // Then notify the waiter.
+  _wait_helper.notify();
+}
+
+WaitHelper::WaitHelper() : _should_wait(true), _monitor(MonitorSupply::reserve()) {
+  if (TraceGCTaskManager) {
+    tty->print_cr("[" INTPTR_FORMAT "]"
+                  " WaitHelper::WaitHelper()"
+                  "  monitor: " INTPTR_FORMAT,
+                  p2i(this), p2i(monitor()));
+  }
+}
+
+void WaitHelper::release_monitor() {
+  assert(_monitor != NULL, "");
+  MonitorSupply::release(_monitor);
+  _monitor = NULL;
+}
+
+WaitHelper::~WaitHelper() {
+  release_monitor();
+}
+
+void WaitHelper::wait_for(bool reset) {
+  if (TraceGCTaskManager) {
+    tty->print_cr("[" INTPTR_FORMAT "]"
+                  " WaitForBarrierGCTask::wait_for()"
+      "  should_wait: %s",
+      p2i(this), should_wait() ? "true" : "false");
+  }
+  {
+    // Grab the lock and check again.
+    MutexLockerEx ml(monitor(), Mutex::_no_safepoint_check_flag);
+    while (should_wait()) {
+      if (TraceGCTaskManager) {
+        tty->print_cr("[" INTPTR_FORMAT "]"
+                      " WaitForBarrierGCTask::wait_for()"
+          "  [" INTPTR_FORMAT "] (%s)->wait()",
+          p2i(this), p2i(monitor()), monitor()->name());
+      }
+      monitor()->wait(Mutex::_no_safepoint_check_flag, 0);
+    }
+    // Reset the flag in case someone reuses this task.
+    if (reset) {
+      set_should_wait(true);
+    }
+    if (TraceGCTaskManager) {
+      tty->print_cr("[" INTPTR_FORMAT "]"
+                    " WaitForBarrierGCTask::wait_for() returns"
+        "  should_wait: %s",
+        p2i(this), should_wait() ? "true" : "false");
+    }
+    // Release monitor().
+  }
+}
+
+void WaitHelper::notify() {
+  MutexLockerEx ml(monitor(), Mutex::_no_safepoint_check_flag);
+  set_should_wait(false);
+  // Waiter doesn't miss the notify in the wait_for method
+  // since it checks the flag after grabbing the monitor.
+  if (TraceGCTaskManager) {
+    tty->print_cr("[" INTPTR_FORMAT "]"
+                  " WaitForBarrierGCTask::do_it()"
+                  "  [" INTPTR_FORMAT "] (%s)->notify_all()",
+                  p2i(this), p2i(monitor()), monitor()->name());
+  }
+  monitor()->notify_all();
+}
+
+Mutex*                   MonitorSupply::_lock     = NULL;
+GrowableArray<Monitor*>* MonitorSupply::_freelist = NULL;
+
+Monitor* MonitorSupply::reserve() {
+  Monitor* result = NULL;
+  // Lazy initialization: possible race.
+  if (lock() == NULL) {
+    _lock = new Mutex(Mutex::barrier,                  // rank
+                      "MonitorSupply mutex",           // name
+                      Mutex::_allow_vm_block_flag);    // allow_vm_block
+  }
+  {
+    MutexLockerEx ml(lock());
+    // Lazy initialization.
+    if (freelist() == NULL) {
+      _freelist =
+        new(ResourceObj::C_HEAP, mtGC) GrowableArray<Monitor*>(ParallelGCThreads,
+                                                         true);
+    }
+    if (! freelist()->is_empty()) {
+      result = freelist()->pop();
+    } else {
+      result = new Monitor(Mutex::barrier,                  // rank
+                           "MonitorSupply monitor",         // name
+                           Mutex::_allow_vm_block_flag,     // allow_vm_block
+                           Monitor::_safepoint_check_never);
+    }
+    guarantee(result != NULL, "shouldn't return NULL");
+    assert(!result->is_locked(), "shouldn't be locked");
+    // release lock().
+  }
+  return result;
+}
+
+void MonitorSupply::release(Monitor* instance) {
+  assert(instance != NULL, "shouldn't release NULL");
+  assert(!instance->is_locked(), "shouldn't be locked");
+  {
+    MutexLockerEx ml(lock());
+    freelist()->push(instance);
+    // release lock().
+  }
+}