/hg/icedtea7-forest/hotspot: 3 new changesets

enevill at icedtea.classpath.org enevill at icedtea.classpath.org
Tue Dec 15 09:00:30 UTC 2015


changeset 5b6efbae9fea in /hg/icedtea7-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea7-forest/hotspot?cmd=changeset;node=5b6efbae9fea
author: aph
date: Wed Nov 04 13:38:38 2015 +0100

	8138966: Intermittent SEGV running ParallelGC
	Summary: Add necessary memory fences so that the parallel threads are unable to observe partially filled block tables.
	Reviewed-by: tschatzl


changeset c7679d143590 in /hg/icedtea7-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea7-forest/hotspot?cmd=changeset;node=c7679d143590
author: enevill
date: Thu Nov 19 15:15:20 2015 +0000

	8143067: aarch64: guarantee failure in javac
	Summary: Fix adrp going out of range during code relocation
	Reviewed-by: aph, kvn


changeset eeb4a3ec4563 in /hg/icedtea7-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea7-forest/hotspot?cmd=changeset;node=eeb4a3ec4563
author: hshi
date: Tue Nov 24 09:02:26 2015 +0000

	8143285: aarch64: Missing load acquire when checking if ConstantPoolCacheEntry is resolved
	Reviewed-by: roland, aph


diffstat:

 src/cpu/aarch64/vm/assembler_aarch64.cpp                              |  101 +++++----
 src/cpu/aarch64/vm/interp_masm_aarch64.cpp                            |    5 +-
 src/share/vm/gc_implementation/parallelScavenge/psParallelCompact.hpp |    7 +-
 3 files changed, 59 insertions(+), 54 deletions(-)

diffs (208 lines):

diff -r 3f07f893bb63 -r eeb4a3ec4563 src/cpu/aarch64/vm/assembler_aarch64.cpp
--- a/src/cpu/aarch64/vm/assembler_aarch64.cpp	Fri Nov 20 00:10:48 2015 +0000
+++ b/src/cpu/aarch64/vm/assembler_aarch64.cpp	Tue Nov 24 09:02:26 2015 +0000
@@ -1597,24 +1597,20 @@
       unsigned offset_lo = dest & 0xfff;
       offset = adr_page - pc_page;
 
-      // We handle 3 types of PC relative addressing
+      // We handle 4 types of PC relative addressing
       //   1 - adrp    Rx, target_page
       //       ldr/str Ry, [Rx, #offset_in_page]
       //   2 - adrp    Rx, target_page
       //       add     Ry, Rx, #offset_in_page
       //   3 - adrp    Rx, target_page (page aligned reloc, offset == 0)
+      //       movk    Rx, #imm16<<32
+      //   4 - adrp    Rx, target_page (page aligned reloc, offset == 0)
       //
-      // In the first 2 cases we must check that Rx is the same in the
-      // adrp and the subsequent ldr/str or add instruction. Otherwise
-      // we could accidentally end up treating a type 3 relocation as
-      // a type 1 or 2 just because it happened to be followed by a
-      // random unrelated ldr/str or add instruction.
+      // In the first 3 cases we must check that Rx is the same in the adrp and the
+      // subsequent ldr/str, add or movk instruction. Otherwise we could accidentally end
+      // up treating a type 4 relocation as a type 1, 2 or 3 just because it happened
+      // to be followed by a random unrelated ldr/str, add or movk instruction.
       //
-      // In the case of a type 3 relocation, we know that these are
-      // only generated for the safepoint polling page, the crc table
-      // base or the card type byte map base so we assert as much
-      // and of course that the offset is 0.
-      // 
       // In jdk7 the card type byte map base is aligned on a 1K
       // boundary which may fail to be 4K aligned. In that case the
       // card table load will fall into category 2.
@@ -1628,24 +1624,21 @@
 	Instruction_aarch64::patch(branch + sizeof (unsigned),
 				    21, 10, offset_lo >> size);
 	guarantee(((dest >> size) << size) == dest, "misaligned target");
-        instructions = 2;
+	instructions = 2;
       } else if (Instruction_aarch64::extract(insn2, 31, 22) == 0b1001000100 &&
 		Instruction_aarch64::extract(insn, 4, 0) ==
 			Instruction_aarch64::extract(insn2, 4, 0)) {
 	// add (immediate)
-	assert (((jbyte *)target !=
-		 ((CardTableModRefBS*)(Universe::heap()->barrier_set()))->byte_map_base) ||
-		(offset_lo & 0x3FFl) == 0, "offset must be 0x400 aligned for crc_table");
 	Instruction_aarch64::patch(branch + sizeof (unsigned),
 				   21, 10, offset_lo);
+	instructions = 2;
+      } else if (Instruction_aarch64::extract(insn2, 31, 21) == 0b11110010110 &&
+                   Instruction_aarch64::extract(insn, 4, 0) ==
+                     Instruction_aarch64::extract(insn2, 4, 0)) {
+        // movk #imm16<<32
+        Instruction_aarch64::patch(branch + 4, 20, 5, (uint64_t)target >> 32);
+        offset &= (1<<20)-1;
         instructions = 2;
-      } else {
-	assert((jbyte *)target ==
-		((CardTableModRefBS*)(Universe::heap()->barrier_set()))->byte_map_base ||
-               target == StubRoutines::crc_table_addr() ||
-               (address)target == os::get_polling_page(),
-	       "adrp must be polling page, crc_table or byte map base");
-	assert(offset_lo == 0, "offset must be 0 for polling page, crc_table or byte map base");
       }
     }
     int offset_lo = offset & 3;
@@ -1729,20 +1722,16 @@
       // Return the target address for the following sequences
       //   1 - adrp    Rx, target_page
       //       ldr/str Ry, [Rx, #offset_in_page]
-      //   [ 2 - adrp    Rx, target_page         ] Not handled
-      //   [    add     Ry, Rx, #offset_in_page  ]
+      //   2 - adrp    Rx, target_page
+      //       add     Ry, Rx, #offset_in_page
       //   3 - adrp    Rx, target_page (page aligned reloc, offset == 0)
+      //       movk    Rx, #imm12<<32
+      //   4 - adrp    Rx, target_page (page aligned reloc, offset == 0)
       //
-      // In the case of type 1 we check that the register is the same and
+      // In the first two cases  we check that the register is the same and
       // return the target_page + the offset within the page.
-      //
       // Otherwise we assume it is a page aligned relocation and return
-      // the target page only. The only cases this is generated is for
-      // the safepoint polling page or for the card table byte map base so
-      // we assert as much.
-      //
-      // Note: Strangely, we do not handle 'type 2' relocation (adrp followed
-      // by add) which is handled in pd_patch_instruction above.
+      // the target page only.
       //
       unsigned insn2 = ((unsigned*)insn_addr)[1];
       if (Instruction_aarch64::extract(insn2, 29, 24) == 0b111001 &&
@@ -1752,11 +1741,19 @@
 	unsigned int byte_offset = Instruction_aarch64::extract(insn2, 21, 10);
 	unsigned int size = Instruction_aarch64::extract(insn2, 31, 30);
 	return address(target_page + (byte_offset << size));
+      } else if (Instruction_aarch64::extract(insn2, 31, 22) == 0b1001000100 &&
+                Instruction_aarch64::extract(insn, 4, 0) ==
+                        Instruction_aarch64::extract(insn2, 4, 0)) {
+        // add (immediate)
+        unsigned int byte_offset = Instruction_aarch64::extract(insn2, 21, 10);
+        return address(target_page + byte_offset);
       } else {
-	assert((jbyte *)target_page ==
-		((CardTableModRefBS*)(Universe::heap()->barrier_set()))->byte_map_base ||
-               (address)target_page == os::get_polling_page(),
-	       "adrp must be polling page or byte map base");
+        if (Instruction_aarch64::extract(insn2, 31, 21) == 0b11110010110  &&
+               Instruction_aarch64::extract(insn, 4, 0) ==
+                 Instruction_aarch64::extract(insn2, 4, 0)) {
+          target_page = (target_page & 0xffffffff) |
+                         ((uint64_t)Instruction_aarch64::extract(insn2, 20, 5) << 32);
+        }
 	return (address)target_page;
       }
     } else {
@@ -4776,22 +4773,26 @@
 
 void MacroAssembler::adrp(Register reg1, const Address &dest, unsigned long &byte_offset) {
   relocInfo::relocType rtype = dest.rspec().reloc()->type();
-  if (uabs(pc() - dest.target()) >= (1LL << 32)) {
-    guarantee(rtype == relocInfo::none
-	      || rtype == relocInfo::external_word_type
-	      || rtype == relocInfo::poll_type
-	      || rtype == relocInfo::poll_return_type,
-	      "can only use a fixed address with an ADRP");
-    // Out of range.  This doesn't happen very often, but we have to
-    // handle it
-    mov(reg1, dest);
-    byte_offset = 0;
+  unsigned long low_page = (unsigned long)CodeCache::low_bound() >> 12;
+  unsigned long high_page = (unsigned long)(CodeCache::high_bound()-1) >> 12;
+  unsigned long dest_page = (unsigned long)dest.target() >> 12;
+  long offset_low = dest_page - low_page;
+  long offset_high = dest_page - high_page;
+
+  InstructionMark im(this);
+  code_section()->relocate(inst_mark(), dest.rspec());
+  // 8143067: Ensure that the adrp can reach the dest from anywhere within
+  // the code cache so that if it is relocated we know it will still reach
+  if (offset_high >= -(1<<20) && offset_low < (1<<20)) {
+    _adrp(reg1, dest.target());
   } else {
-    InstructionMark im(this);
-    code_section()->relocate(inst_mark(), dest.rspec());
-    byte_offset = (uint64_t)dest.target() & 0xfff;
-    _adrp(reg1, dest.target());
+    unsigned long pc_page = (unsigned long)pc() >> 12;
+    long offset = dest_page - pc_page;
+    offset = (offset & ((1<<20)-1)) << 12;
+    _adrp(reg1, pc()+offset);
+    movk(reg1, ((unsigned long)dest.target() >> 32) & 0xffff, 32);
   }
+  byte_offset = (unsigned long)dest.target() & 0xfff;
 }
 
 void MacroAssembler::build_frame(int framesize) {
diff -r 3f07f893bb63 -r eeb4a3ec4563 src/cpu/aarch64/vm/interp_masm_aarch64.cpp
--- a/src/cpu/aarch64/vm/interp_masm_aarch64.cpp	Fri Nov 20 00:10:48 2015 +0000
+++ b/src/cpu/aarch64/vm/interp_masm_aarch64.cpp	Tue Nov 24 09:02:26 2015 +0000
@@ -197,10 +197,11 @@
   get_cache_and_index_at_bcp(cache, index, bcp_offset, index_size);
   // We use a 32-bit load here since the layout of 64-bit words on
   // little-endian machines allow us that.
-  // n.b. unlike x86 cache alreeady includes the index offset
-  ldrw(bytecode, Address(cache,
+  // n.b. unlike x86 cache already includes the index offset
+  lea(bytecode, Address(cache,
 			 constantPoolCacheOopDesc::base_offset()
 			 + ConstantPoolCacheEntry::indices_offset()));
+  ldarw(bytecode, bytecode);
   const int shift_count = (1 + byte_no) * BitsPerByte;
   ubfx(bytecode, bytecode, shift_count, BitsPerByte);
 }
diff -r 3f07f893bb63 -r eeb4a3ec4563 src/share/vm/gc_implementation/parallelScavenge/psParallelCompact.hpp
--- a/src/share/vm/gc_implementation/parallelScavenge/psParallelCompact.hpp	Fri Nov 20 00:10:48 2015 +0000
+++ b/src/share/vm/gc_implementation/parallelScavenge/psParallelCompact.hpp	Tue Nov 24 09:02:26 2015 +0000
@@ -349,7 +349,7 @@
     HeapWord*            _partial_obj_addr;
     region_sz_t          _partial_obj_size;
     region_sz_t volatile _dc_and_los;
-    bool                 _blocks_filled;
+    bool        volatile _blocks_filled;
 
 #ifdef ASSERT
     size_t               _blocks_filled_count;   // Number of block table fills.
@@ -503,7 +503,9 @@
 inline bool
 ParallelCompactData::RegionData::blocks_filled() const
 {
-  return _blocks_filled;
+  bool result = _blocks_filled;
+  OrderAccess::acquire();
+  return result;
 }
 
 #ifdef ASSERT
@@ -517,6 +519,7 @@
 inline void
 ParallelCompactData::RegionData::set_blocks_filled()
 {
+  OrderAccess::release();
   _blocks_filled = true;
   // Debug builds count the number of times the table was filled.
   DEBUG_ONLY(Atomic::inc_ptr(&_blocks_filled_count));


More information about the distro-pkg-dev mailing list