--- a/src/hotspot/cpu/aarch64/templateInterpreterGenerator_aarch64.cpp Wed Nov 06 09:50:53 2019 -0500
+++ b/src/hotspot/cpu/aarch64/templateInterpreterGenerator_aarch64.cpp Wed Nov 06 16:52:49 2019 +0100
@@ -285,7 +285,6 @@
}
break;
case Interpreter::java_lang_math_pow :
- fpargs = 2;
if (StubRoutines::dpow() == NULL) {
fn = CAST_FROM_FN_PTR(address, SharedRuntime::dpow);
} else {
--- a/src/hotspot/os_cpu/linux_aarch64/orderAccess_linux_aarch64.hpp Wed Nov 06 09:50:53 2019 -0500
+++ b/src/hotspot/os_cpu/linux_aarch64/orderAccess_linux_aarch64.hpp Wed Nov 06 16:52:49 2019 +0100
@@ -55,14 +55,14 @@
struct OrderAccess::PlatformOrderedLoad<byte_size, X_ACQUIRE>
{
template <typename T>
- T operator()(const volatile T* p) const { T data; __atomic_load(p, &data, __ATOMIC_ACQUIRE); return data; }
+ T operator()(const volatile T* p) const { T data; __atomic_load(const_cast<T*>(p), &data, __ATOMIC_ACQUIRE); return data; }
};
template<size_t byte_size>
struct OrderAccess::PlatformOrderedStore<byte_size, RELEASE_X>
{
template <typename T>
- void operator()(T v, volatile T* p) const { __atomic_store(p, &v, __ATOMIC_RELEASE); }
+ void operator()(T v, volatile T* p) const { __atomic_store(const_cast<T*>(p), &v, __ATOMIC_RELEASE); }
};
template<size_t byte_size>