hotspot/src/cpu/x86/vm/bytes_x86.hpp
changeset 46504 38048d4d20e7
parent 40010 e32d5e545789
child 46619 a3919f5e8d2b
--- a/hotspot/src/cpu/x86/vm/bytes_x86.hpp	Fri May 26 09:48:46 2017 -0400
+++ b/hotspot/src/cpu/x86/vm/bytes_x86.hpp	Fri May 26 13:47:33 2017 -0700
@@ -36,36 +36,85 @@
 #endif // AMD64
 
  public:
-  // Returns true if the byte ordering used by Java is different from the native byte ordering
-  // of the underlying machine. For example, this is true for Intel x86, but false for Solaris
-  // on Sparc.
-  static inline bool is_Java_byte_ordering_different(){ return true; }
+  // Efficient reading and writing of unaligned unsigned data in platform-specific byte ordering
+  template <typename T>
+  static inline T get_native(const void* p) {
+    assert(p != NULL, "null pointer");
+
+    T x;
 
+    if (is_ptr_aligned(p, sizeof(T))) {
+      x = *(T*)p;
+    } else {
+      memcpy(&x, p, sizeof(T));
+    }
+
+    return x;
+  }
 
-  // Efficient reading and writing of unaligned unsigned data in platform-specific byte ordering
-  // (no special code is needed since x86 CPUs can access unaligned data)
-  static inline u2   get_native_u2(address p)         { return *(u2*)p; }
-  static inline u4   get_native_u4(address p)         { return *(u4*)p; }
-  static inline u8   get_native_u8(address p)         { return *(u8*)p; }
+  template <typename T>
+  static inline void put_native(void* p, T x) {
+    assert(p != NULL, "null pointer");
 
-  static inline void put_native_u2(address p, u2 x)   { *(u2*)p = x; }
-  static inline void put_native_u4(address p, u4 x)   { *(u4*)p = x; }
-  static inline void put_native_u8(address p, u8 x)   { *(u8*)p = x; }
+    if (is_ptr_aligned(p, sizeof(T))) {
+      *(T*)p = x;
+    } else {
+      memcpy(p, &x, sizeof(T));
+    }
+  }
 
+  static inline u2   get_native_u2(address p)         { return get_native<u2>((void*)p); }
+  static inline u4   get_native_u4(address p)         { return get_native<u4>((void*)p); }
+  static inline u8   get_native_u8(address p)         { return get_native<u8>((void*)p); }
+  static inline void put_native_u2(address p, u2 x)   { put_native<u2>((void*)p, x); }
+  static inline void put_native_u4(address p, u4 x)   { put_native<u4>((void*)p, x); }
+  static inline void put_native_u8(address p, u8 x)   { put_native<u8>((void*)p, x); }
 
   // Efficient reading and writing of unaligned unsigned data in Java
   // byte ordering (i.e. big-endian ordering). Byte-order reversal is
   // needed since x86 CPUs use little-endian format.
-  static inline u2   get_Java_u2(address p)           { return swap_u2(get_native_u2(p)); }
-  static inline u4   get_Java_u4(address p)           { return swap_u4(get_native_u4(p)); }
-  static inline u8   get_Java_u8(address p)           { return swap_u8(get_native_u8(p)); }
+  template <typename T>
+  static inline T get_Java(const address p) {
+    T x = get_native<T>(p);
+
+    if (Endian::is_Java_byte_ordering_different()) {
+      x = swap<T>(x);
+    }
+
+    return x;
+  }
 
-  static inline void put_Java_u2(address p, u2 x)     { put_native_u2(p, swap_u2(x)); }
-  static inline void put_Java_u4(address p, u4 x)     { put_native_u4(p, swap_u4(x)); }
-  static inline void put_Java_u8(address p, u8 x)     { put_native_u8(p, swap_u8(x)); }
+  template <typename T>
+  static inline void put_Java(address p, T x) {
+    if (Endian::is_Java_byte_ordering_different()) {
+      x = swap<T>(x);
+    }
 
+    put_native<T>(p, x);
+  }
+
+  static inline u2   get_Java_u2(address p)           { return get_Java<u2>(p); }
+  static inline u4   get_Java_u4(address p)           { return get_Java<u4>(p); }
+  static inline u8   get_Java_u8(address p)           { return get_Java<u8>(p); }
+
+  static inline void put_Java_u2(address p, u2 x)     { put_Java<u2>(p, x); }
+  static inline void put_Java_u4(address p, u4 x)     { put_Java<u4>(p, x); }
+  static inline void put_Java_u8(address p, u8 x)     { put_Java<u8>(p, x); }
 
   // Efficient swapping of byte ordering
+  template <typename T>
+  static T swap(T x) {
+    switch (sizeof(T)) {
+    case sizeof(u1): return x;
+    case sizeof(u2): return swap_u2(x);
+    case sizeof(u4): return swap_u4(x);
+    case sizeof(u8): return swap_u8(x);
+    default:
+      guarantee(false, "invalid size: " SIZE_FORMAT "\n", sizeof(T));
+      return 0;
+    }
+  }
+
   static inline u2   swap_u2(u2 x);                   // compiler-dependent implementation
   static inline u4   swap_u4(u4 x);                   // compiler-dependent implementation
   static inline u8   swap_u8(u8 x);