src/hotspot/share/jfr/periodic/jfrThreadCPULoadEvent.cpp
changeset 50113 caf115bb98ad
child 57360 5d043a159d5c
--- /dev/null	Thu Jan 01 00:00:00 1970 +0000
+++ b/src/hotspot/share/jfr/periodic/jfrThreadCPULoadEvent.cpp	Tue May 15 20:24:34 2018 +0200
@@ -0,0 +1,145 @@
+/*
+ * Copyright (c) 2017, 2018, Oracle and/or its affiliates. All rights reserved.
+ * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
+ *
+ * This code is free software; you can redistribute it and/or modify it
+ * under the terms of the GNU General Public License version 2 only, as
+ * published by the Free Software Foundation.
+ *
+ * This code is distributed in the hope that it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
+ * version 2 for more details (a copy is included in the LICENSE file that
+ * accompanied this code).
+ *
+ * You should have received a copy of the GNU General Public License version
+ * 2 along with this work; if not, write to the Free Software Foundation,
+ * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
+ *
+ * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
+ * or visit www.oracle.com if you need additional information or have any
+ * questions.
+ *
+ */
+
+#include "precompiled.hpp"
+#include "logging/log.hpp"
+#include "jfr/jfrEvents.hpp"
+#include "jfr/periodic/jfrThreadCPULoadEvent.hpp"
+#include "jfr/support/jfrThreadId.hpp"
+#include "jfr/support/jfrThreadLocal.hpp"
+#include "jfr/utilities/jfrTime.hpp"
+#include "utilities/globalDefinitions.hpp"
+#include "runtime/os.hpp"
+#include "runtime/thread.inline.hpp"
+#include "runtime/threadSMR.inline.hpp"
+
+jlong JfrThreadCPULoadEvent::get_wallclock_time() {
+  return os::javaTimeNanos();
+}
+
+int JfrThreadCPULoadEvent::_last_active_processor_count = 0;
+
+int JfrThreadCPULoadEvent::get_processor_count() {
+  int cur_processor_count = os::active_processor_count();
+  int last_processor_count = _last_active_processor_count;
+  _last_active_processor_count = cur_processor_count;
+
+  // If the number of processors decreases, we don't know at what point during
+  // the sample interval this happened, so use the largest number to try
+  // to avoid percentages above 100%
+  return MAX2(cur_processor_count, last_processor_count);
+}
+
+// Returns false if the thread has not been scheduled since the last call to updateEvent
+// (i.e. the delta for both system and user time is 0 milliseconds)
+bool JfrThreadCPULoadEvent::update_event(EventThreadCPULoad& event, JavaThread* thread, jlong cur_wallclock_time, int processor_count) {
+  JfrThreadLocal* const tl = thread->jfr_thread_local();
+
+  jlong cur_cpu_time = os::thread_cpu_time(thread, true);
+  jlong prev_cpu_time = tl->get_cpu_time();
+
+  jlong prev_wallclock_time = tl->get_wallclock_time();
+  tl->set_wallclock_time(cur_wallclock_time);
+
+  // Threshold of 1 ms
+  if (cur_cpu_time - prev_cpu_time < 1 * NANOSECS_PER_MILLISEC) {
+    return false;
+  }
+
+  jlong cur_user_time = os::thread_cpu_time(thread, false);
+  jlong prev_user_time = tl->get_user_time();
+
+  jlong cur_system_time = cur_cpu_time - cur_user_time;
+  jlong prev_system_time = prev_cpu_time - prev_user_time;
+
+  // The user and total cpu usage clocks can have different resolutions, which can
+  // make us see decreasing system time. Ensure time doesn't go backwards.
+  if (prev_system_time > cur_system_time) {
+    cur_cpu_time += prev_system_time - cur_system_time;
+    cur_system_time = prev_system_time;
+  }
+
+  jlong user_time = cur_user_time - prev_user_time;
+  jlong system_time = cur_system_time - prev_system_time;
+  jlong wallclock_time = cur_wallclock_time - prev_wallclock_time;
+  jlong total_available_time = wallclock_time * processor_count;
+
+  // Avoid reporting percentages above the theoretical max
+  if (user_time + system_time > wallclock_time) {
+    jlong excess = user_time + system_time - wallclock_time;
+    if (user_time > excess) {
+      user_time -= excess;
+      cur_user_time -= excess;
+      cur_cpu_time -= excess;
+    } else {
+      cur_cpu_time -= excess;
+      excess -= user_time;
+      user_time = 0;
+      cur_user_time = 0;
+      system_time -= excess;
+    }
+  }
+  event.set_user(total_available_time > 0 ? (double)user_time / total_available_time : 0);
+  event.set_system(total_available_time > 0 ? (double)system_time / total_available_time : 0);
+  tl->set_user_time(cur_user_time);
+  tl->set_cpu_time(cur_cpu_time);
+  return true;
+}
+
+void JfrThreadCPULoadEvent::send_events() {
+  Thread* periodic_thread = Thread::current();
+  JfrThreadLocal* const periodic_thread_tl = periodic_thread->jfr_thread_local();
+  traceid periodic_thread_id = periodic_thread_tl->thread_id();
+  const int processor_count = JfrThreadCPULoadEvent::get_processor_count();
+  JfrTicks event_time = JfrTicks::now();
+  jlong cur_wallclock_time = JfrThreadCPULoadEvent::get_wallclock_time();
+
+  JavaThreadIteratorWithHandle jtiwh;
+  while (JavaThread* jt = jtiwh.next()) {
+    EventThreadCPULoad event(UNTIMED);
+    if (JfrThreadCPULoadEvent::update_event(event, jt, cur_wallclock_time, processor_count)) {
+      event.set_starttime(event_time);
+      if (jt != periodic_thread) {
+        // Commit reads the thread id from this thread's trace data, so put it there temporarily
+        periodic_thread_tl->set_thread_id(JFR_THREAD_ID(jt));
+      } else {
+        periodic_thread_tl->set_thread_id(periodic_thread_id);
+      }
+      event.commit();
+    }
+  }
+  log_trace(jfr)("Measured CPU usage for %d threads in %.3f milliseconds", jtiwh.length(),
+    (double)(JfrTicks::now() - event_time).milliseconds());
+  // Restore this thread's thread id
+  periodic_thread_tl->set_thread_id(periodic_thread_id);
+}
+
+void JfrThreadCPULoadEvent::send_event_for_thread(JavaThread* jt) {
+  EventThreadCPULoad event;
+  if (event.should_commit()) {
+    if (update_event(event, jt, get_wallclock_time(), get_processor_count())) {
+      event.commit();
+    }
+  }
+}