/hg/icedtea7-forest/hotspot: 2 new changesets

andrew at icedtea.classpath.org andrew at icedtea.classpath.org
Tue Jun 14 18:13:36 UTC 2016


changeset a136b8a1ad7a in /hg/icedtea7-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea7-forest/hotspot?cmd=changeset;node=a136b8a1ad7a
author: minqi
date: Wed Jun 01 19:31:26 2016 +0100

	2178143, PR2958: JVM crashes if the number of bound CPUs changed during runtime
	Summary: Supply a new flag -XX:+AssumeMP to workaround the problem. With the flag is turned on, assume VM run on MP platform so is_MP() will return true that sync calls will not skip away.
	Reviewed-by: dholmes, acorn, dcubed, jmasa
	Contributed-by: yumin.qi at oracle.com


changeset a778398a0cdb in /hg/icedtea7-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea7-forest/hotspot?cmd=changeset;node=a778398a0cdb
author: andrew
date: Fri Jun 10 17:51:24 2016 +0100

	Merge


diffstat:

 src/cpu/aarch64/vm/aarch64.ad                |  14 +++++-----
 src/cpu/aarch64/vm/assembler_aarch64.cpp     |  16 +++++++----
 src/cpu/aarch64/vm/globals_aarch64.hpp       |   4 --
 src/cpu/aarch64/vm/stubGenerator_aarch64.cpp |  39 ++++++++++++++++-----------
 src/cpu/aarch64/vm/templateTable_aarch64.cpp |   4 +-
 src/share/vm/runtime/arguments.cpp           |   8 +++++
 src/share/vm/runtime/globals.hpp             |   3 ++
 src/share/vm/runtime/os.hpp                  |   2 +-
 8 files changed, 54 insertions(+), 36 deletions(-)

diffs (278 lines):

diff -r 2d8e12787f80 -r a778398a0cdb src/cpu/aarch64/vm/aarch64.ad
--- a/src/cpu/aarch64/vm/aarch64.ad	Tue Apr 19 19:52:39 2016 -0700
+++ b/src/cpu/aarch64/vm/aarch64.ad	Fri Jun 10 17:51:24 2016 +0100
@@ -9090,21 +9090,21 @@
   %}
 %}
 
-instruct rolI_rReg_Var_C_32(iRegLNoSp dst, iRegL src, iRegI shift, immI_32 c_32, rFlagsReg cr)
+instruct rolI_rReg_Var_C_32(iRegINoSp dst, iRegI src, iRegI shift, immI_32 c_32, rFlagsReg cr)
 %{
   match(Set dst (OrI (LShiftI src shift) (URShiftI src (SubI c_32 shift))));
 
   expand %{
-    rolL_rReg(dst, src, shift, cr);
-  %}
-%}
-
-instruct rolI_rReg_Var_C0(iRegLNoSp dst, iRegL src, iRegI shift, immI0 c0, rFlagsReg cr)
+    rolI_rReg(dst, src, shift, cr);
+  %}
+%}
+
+instruct rolI_rReg_Var_C0(iRegINoSp dst, iRegI src, iRegI shift, immI0 c0, rFlagsReg cr)
 %{
   match(Set dst (OrI (LShiftI src shift) (URShiftI src (SubI c0 shift))));
 
   expand %{
-    rolL_rReg(dst, src, shift, cr);
+    rolI_rReg(dst, src, shift, cr);
   %}
 %}
 
diff -r 2d8e12787f80 -r a778398a0cdb src/cpu/aarch64/vm/assembler_aarch64.cpp
--- a/src/cpu/aarch64/vm/assembler_aarch64.cpp	Tue Apr 19 19:52:39 2016 -0700
+++ b/src/cpu/aarch64/vm/assembler_aarch64.cpp	Fri Jun 10 17:51:24 2016 +0100
@@ -1632,7 +1632,10 @@
                      Instruction_aarch64::extract(insn2, 4, 0)) {
         // movk #imm16<<32
         Instruction_aarch64::patch(branch + 4, 20, 5, (uint64_t)target >> 32);
-        offset &= (1<<20)-1;
+        long dest = ((long)target & 0xffffffffL) | ((long)branch & 0xffff00000000L);
+        long pc_page = (long)branch >> 12;
+        long adr_page = (long)dest >> 12;
+        offset = adr_page - pc_page;
         instructions = 2;
       }
     }
@@ -4920,11 +4923,12 @@
   if (offset_high >= -(1<<20) && offset_low < (1<<20)) {
     _adrp(reg1, dest.target());
   } else {
-    unsigned long pc_page = (unsigned long)pc() >> 12;
-    long offset = dest_page - pc_page;
-    offset = (offset & ((1<<20)-1)) << 12;
-    _adrp(reg1, pc()+offset);
-    movk(reg1, ((unsigned long)dest.target() >> 32), 32);
+    unsigned long target = (unsigned long)dest.target();
+    unsigned long adrp_target
+      = (target & 0xffffffffUL) | ((unsigned long)pc() & 0xffff00000000UL);
+           
+    _adrp(reg1, (address)adrp_target);
+    movk(reg1, target >> 32, 32);
   }
   byte_offset = (unsigned long)dest.target() & 0xfff;
 }
diff -r 2d8e12787f80 -r a778398a0cdb src/cpu/aarch64/vm/globals_aarch64.hpp
--- a/src/cpu/aarch64/vm/globals_aarch64.hpp	Tue Apr 19 19:52:39 2016 -0700
+++ b/src/cpu/aarch64/vm/globals_aarch64.hpp	Fri Jun 10 17:51:24 2016 +0100
@@ -48,11 +48,7 @@
 // the the vep is aligned at CodeEntryAlignment whereas c2 only aligns
 // the uep and the vep doesn't get real alignment but just slops on by
 // only assured that the entry instruction meets the 5 byte size requirement.
-#ifdef COMPILER2
 define_pd_global(intx, CodeEntryAlignment,       64);
-#else
-define_pd_global(intx, CodeEntryAlignment,       16);
-#endif // COMPILER2
 define_pd_global(intx, OptoLoopAlignment,        16);
 define_pd_global(intx, InlineFrequencyCount,     100);
 
diff -r 2d8e12787f80 -r a778398a0cdb src/cpu/aarch64/vm/stubGenerator_aarch64.cpp
--- a/src/cpu/aarch64/vm/stubGenerator_aarch64.cpp	Tue Apr 19 19:52:39 2016 -0700
+++ b/src/cpu/aarch64/vm/stubGenerator_aarch64.cpp	Fri Jun 10 17:51:24 2016 +0100
@@ -687,7 +687,7 @@
   //     count   -  element count
   //     tmp     - scratch register
   //
-  //     Destroy no registers!
+  //     Destroy no registers except rscratch1 and rscratch2
   //
   void  gen_write_ref_array_pre_barrier(Register addr, Register count, bool dest_uninitialized) {
     BarrierSet* bs = Universe::heap()->barrier_set();
@@ -696,12 +696,13 @@
     case BarrierSet::G1SATBCTLogging:
       // With G1, don't generate the call if we statically know that the target in uninitialized
       if (!dest_uninitialized) {
-	__ push(RegSet::range(r0, r29), sp);         // integer registers except lr & sp
+	__ push_call_clobbered_registers();
 	if (count == c_rarg0) {
 	  if (addr == c_rarg1) {
 	    // exactly backwards!!
-	    __ stp(c_rarg0, c_rarg1, __ pre(sp, -2 * wordSize));
-	    __ ldp(c_rarg1, c_rarg0, __ post(sp, -2 * wordSize));
+            __ mov(rscratch1, c_rarg0);
+            __ mov(c_rarg0, c_rarg1);
+            __ mov(c_rarg1, rscratch1);
 	  } else {
 	    __ mov(c_rarg1, count);
 	    __ mov(c_rarg0, addr);
@@ -711,7 +712,7 @@
 	  __ mov(c_rarg1, count);
 	}
 	__ call_VM_leaf(CAST_FROM_FN_PTR(address, BarrierSet::static_write_ref_array_pre), 2);
-	__ pop(RegSet::range(r0, r29), sp);         // integer registers except lr & sp        }
+	__ pop_call_clobbered_registers();
 	break;
       case BarrierSet::CardTableModRef:
       case BarrierSet::CardTableExtension:
@@ -742,7 +743,7 @@
       case BarrierSet::G1SATBCTLogging:
 
         {
-	  __ push(RegSet::range(r0, r29), sp);         // integer registers except lr & sp
+	  __ push_call_clobbered_registers();
           // must compute element count unless barrier set interface is changed (other platforms supply count)
           assert_different_registers(start, end, scratch);
           __ lea(scratch, Address(end, BytesPerHeapOop));
@@ -751,7 +752,7 @@
           __ mov(c_rarg0, start);
           __ mov(c_rarg1, scratch);
           __ call_VM_leaf(CAST_FROM_FN_PTR(address, BarrierSet::static_write_ref_array_post), 2);
-	  __ pop(RegSet::range(r0, r29), sp);         // integer registers except lr & sp        }
+	  __ pop_call_clobbered_registers();
         }
         break;
       case BarrierSet::CardTableModRef:
@@ -811,7 +812,7 @@
     assert_different_registers(s, d, count, rscratch1);
 
     Label again, large, small;
-    __ align(6);
+    __ align(CodeEntryAlignment);
     __ bind(start);
     __ cmp(count, 8);
     __ br(Assembler::LO, small);
@@ -856,7 +857,7 @@
 
     __ ret(lr);
 
-    __ align(6);
+    __ align(CodeEntryAlignment);
     __ bind(large);
 
     // Fill 8 registers
@@ -1007,7 +1008,8 @@
       }
       // rscratch2 is the byte adjustment needed to align s.
       __ cbz(rscratch2, aligned);
-      __ lsr(rscratch2, rscratch2, exact_log2(granularity));
+      int shift = exact_log2(granularity);
+      if (shift)  __ lsr(rscratch2, rscratch2, shift);
       __ sub(count, count, rscratch2);
 
 #if 0
@@ -1386,10 +1388,10 @@
   //   no-overlap entry point used by generate_conjoint_long_oop_copy().
   //
   address generate_disjoint_oop_copy(bool aligned, address *entry,
-				     const char *name, bool dest_uninitialized = false) {
+				     const char *name, bool dest_uninitialized) {
     const bool is_oop = true;
     const size_t size = UseCompressedOops ? sizeof (jint) : sizeof (jlong);
-    return generate_disjoint_copy(size, aligned, is_oop, entry, name);
+    return generate_disjoint_copy(size, aligned, is_oop, entry, name, dest_uninitialized);
   }
 
   // Arguments:
@@ -1404,10 +1406,11 @@
   //
   address generate_conjoint_oop_copy(bool aligned,
 				     address nooverlap_target, address *entry,
-				     const char *name, bool dest_uninitialized = false) {
+				     const char *name, bool dest_uninitialized) {
     const bool is_oop = true;
     const size_t size = UseCompressedOops ? sizeof (jint) : sizeof (jlong);
-    return generate_conjoint_copy(size, aligned, is_oop, nooverlap_target, entry, name);
+    return generate_conjoint_copy(size, aligned, is_oop, nooverlap_target, entry,
+                                  name, dest_uninitialized);
   }
 
 
@@ -1514,6 +1517,8 @@
     }
 #endif //ASSERT
 
+    gen_write_ref_array_pre_barrier(to, count, dest_uninitialized);
+
     // save the original count
     __ mov(count_save, count);
 
@@ -1655,9 +1660,11 @@
       bool aligned = !UseCompressedOops;
 
       StubRoutines::_arrayof_oop_disjoint_arraycopy
-	= generate_disjoint_oop_copy(aligned, &entry, "arrayof_oop_disjoint_arraycopy");
+	= generate_disjoint_oop_copy(aligned, &entry, "arrayof_oop_disjoint_arraycopy",
+                                     /*dest_uninitialized*/false);
       StubRoutines::_arrayof_oop_arraycopy
-	= generate_conjoint_oop_copy(aligned, entry, &entry_oop_arraycopy, "arrayof_oop_arraycopy");
+	= generate_conjoint_oop_copy(aligned, entry, &entry_oop_arraycopy, "arrayof_oop_arraycopy",
+                                     /*dest_uninitialized*/false);
       // Aligned versions without pre-barriers
       StubRoutines::_arrayof_oop_disjoint_arraycopy_uninit
 	= generate_disjoint_oop_copy(aligned, &entry, "arrayof_oop_disjoint_arraycopy_uninit",
diff -r 2d8e12787f80 -r a778398a0cdb src/cpu/aarch64/vm/templateTable_aarch64.cpp
--- a/src/cpu/aarch64/vm/templateTable_aarch64.cpp	Tue Apr 19 19:52:39 2016 -0700
+++ b/src/cpu/aarch64/vm/templateTable_aarch64.cpp	Fri Jun 10 17:51:24 2016 +0100
@@ -3032,7 +3032,7 @@
   __ null_check(r0);
   switch (state) {
   case itos:
-    __ ldr(r0, Address(r0, r1, Address::lsl(0)));
+    __ ldrw(r0, Address(r0, r1, Address::lsl(0)));
     break;
   case atos:
     __ load_heap_oop(r0, Address(r0, r1, Address::lsl(0)));
@@ -3052,7 +3052,7 @@
     __ ldrw(r3, Address(r2, in_bytes(constantPoolCacheOopDesc::base_offset() +
 				     ConstantPoolCacheEntry::flags_offset())));
     __ tbz(r3, ConstantPoolCacheEntry::is_volatile_shift, notVolatile);
-    __ membar(MacroAssembler::LoadLoad);
+    __ membar(MacroAssembler::LoadLoad | MacroAssembler::LoadStore);
     __ bind(notVolatile);
   }
 
diff -r 2d8e12787f80 -r a778398a0cdb src/share/vm/runtime/arguments.cpp
--- a/src/share/vm/runtime/arguments.cpp	Tue Apr 19 19:52:39 2016 -0700
+++ b/src/share/vm/runtime/arguments.cpp	Fri Jun 10 17:51:24 2016 +0100
@@ -3473,6 +3473,14 @@
     set_g1_gc_flags();
   }
 
+  if (AssumeMP && !UseSerialGC) {
+    if (FLAG_IS_DEFAULT(ParallelGCThreads) && ParallelGCThreads == 1) {
+      warning("If the number of processors is expected to increase from one, then"
+              " you should configure the number of parallel GC threads appropriately"
+              " using -XX:ParallelGCThreads=N");
+    }
+  }
+
 #ifdef SERIALGC
   assert(verify_serial_gc_flags(), "SerialGC unset");
 #endif // SERIALGC
diff -r 2d8e12787f80 -r a778398a0cdb src/share/vm/runtime/globals.hpp
--- a/src/share/vm/runtime/globals.hpp	Tue Apr 19 19:52:39 2016 -0700
+++ b/src/share/vm/runtime/globals.hpp	Fri Jun 10 17:51:24 2016 +0100
@@ -480,6 +480,9 @@
   lp64_product(intx, ObjectAlignmentInBytes, 8,                             \
           "Default object alignment in bytes, 8 is minimum")                \
                                                                             \
+  product(bool, AssumeMP, false,                                            \
+          "Instruct the VM to assume multiple processors are available")    \
+                                                                            \
   /* UseMembar is theoretically a temp flag used for memory barrier         \
    * removal testing.  It was supposed to be removed before FCS but has     \
    * been re-added (see 6401008) */                                         \
diff -r 2d8e12787f80 -r a778398a0cdb src/share/vm/runtime/os.hpp
--- a/src/share/vm/runtime/os.hpp	Tue Apr 19 19:52:39 2016 -0700
+++ b/src/share/vm/runtime/os.hpp	Fri Jun 10 17:51:24 2016 +0100
@@ -198,7 +198,7 @@
   // Interface for detecting multiprocessor system
   static inline bool is_MP() {
     assert(_processor_count > 0, "invalid processor count");
-    return _processor_count > 1;
+    return _processor_count > 1 || AssumeMP;
   }
   static julong available_memory();
   static julong physical_memory();


More information about the distro-pkg-dev mailing list