/hg/icedtea7-forest/hotspot: 2 new changesets
andrew at icedtea.classpath.org
andrew at icedtea.classpath.org
Tue Jun 14 18:13:36 UTC 2016
changeset a136b8a1ad7a in /hg/icedtea7-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea7-forest/hotspot?cmd=changeset;node=a136b8a1ad7a
author: minqi
date: Wed Jun 01 19:31:26 2016 +0100
2178143, PR2958: JVM crashes if the number of bound CPUs changed during runtime
Summary: Supply a new flag -XX:+AssumeMP to workaround the problem. With the flag is turned on, assume VM run on MP platform so is_MP() will return true that sync calls will not skip away.
Reviewed-by: dholmes, acorn, dcubed, jmasa
Contributed-by: yumin.qi at oracle.com
changeset a778398a0cdb in /hg/icedtea7-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea7-forest/hotspot?cmd=changeset;node=a778398a0cdb
author: andrew
date: Fri Jun 10 17:51:24 2016 +0100
Merge
diffstat:
src/cpu/aarch64/vm/aarch64.ad | 14 +++++-----
src/cpu/aarch64/vm/assembler_aarch64.cpp | 16 +++++++----
src/cpu/aarch64/vm/globals_aarch64.hpp | 4 --
src/cpu/aarch64/vm/stubGenerator_aarch64.cpp | 39 ++++++++++++++++-----------
src/cpu/aarch64/vm/templateTable_aarch64.cpp | 4 +-
src/share/vm/runtime/arguments.cpp | 8 +++++
src/share/vm/runtime/globals.hpp | 3 ++
src/share/vm/runtime/os.hpp | 2 +-
8 files changed, 54 insertions(+), 36 deletions(-)
diffs (278 lines):
diff -r 2d8e12787f80 -r a778398a0cdb src/cpu/aarch64/vm/aarch64.ad
--- a/src/cpu/aarch64/vm/aarch64.ad Tue Apr 19 19:52:39 2016 -0700
+++ b/src/cpu/aarch64/vm/aarch64.ad Fri Jun 10 17:51:24 2016 +0100
@@ -9090,21 +9090,21 @@
%}
%}
-instruct rolI_rReg_Var_C_32(iRegLNoSp dst, iRegL src, iRegI shift, immI_32 c_32, rFlagsReg cr)
+instruct rolI_rReg_Var_C_32(iRegINoSp dst, iRegI src, iRegI shift, immI_32 c_32, rFlagsReg cr)
%{
match(Set dst (OrI (LShiftI src shift) (URShiftI src (SubI c_32 shift))));
expand %{
- rolL_rReg(dst, src, shift, cr);
- %}
-%}
-
-instruct rolI_rReg_Var_C0(iRegLNoSp dst, iRegL src, iRegI shift, immI0 c0, rFlagsReg cr)
+ rolI_rReg(dst, src, shift, cr);
+ %}
+%}
+
+instruct rolI_rReg_Var_C0(iRegINoSp dst, iRegI src, iRegI shift, immI0 c0, rFlagsReg cr)
%{
match(Set dst (OrI (LShiftI src shift) (URShiftI src (SubI c0 shift))));
expand %{
- rolL_rReg(dst, src, shift, cr);
+ rolI_rReg(dst, src, shift, cr);
%}
%}
diff -r 2d8e12787f80 -r a778398a0cdb src/cpu/aarch64/vm/assembler_aarch64.cpp
--- a/src/cpu/aarch64/vm/assembler_aarch64.cpp Tue Apr 19 19:52:39 2016 -0700
+++ b/src/cpu/aarch64/vm/assembler_aarch64.cpp Fri Jun 10 17:51:24 2016 +0100
@@ -1632,7 +1632,10 @@
Instruction_aarch64::extract(insn2, 4, 0)) {
// movk #imm16<<32
Instruction_aarch64::patch(branch + 4, 20, 5, (uint64_t)target >> 32);
- offset &= (1<<20)-1;
+ long dest = ((long)target & 0xffffffffL) | ((long)branch & 0xffff00000000L);
+ long pc_page = (long)branch >> 12;
+ long adr_page = (long)dest >> 12;
+ offset = adr_page - pc_page;
instructions = 2;
}
}
@@ -4920,11 +4923,12 @@
if (offset_high >= -(1<<20) && offset_low < (1<<20)) {
_adrp(reg1, dest.target());
} else {
- unsigned long pc_page = (unsigned long)pc() >> 12;
- long offset = dest_page - pc_page;
- offset = (offset & ((1<<20)-1)) << 12;
- _adrp(reg1, pc()+offset);
- movk(reg1, ((unsigned long)dest.target() >> 32), 32);
+ unsigned long target = (unsigned long)dest.target();
+ unsigned long adrp_target
+ = (target & 0xffffffffUL) | ((unsigned long)pc() & 0xffff00000000UL);
+
+ _adrp(reg1, (address)adrp_target);
+ movk(reg1, target >> 32, 32);
}
byte_offset = (unsigned long)dest.target() & 0xfff;
}
diff -r 2d8e12787f80 -r a778398a0cdb src/cpu/aarch64/vm/globals_aarch64.hpp
--- a/src/cpu/aarch64/vm/globals_aarch64.hpp Tue Apr 19 19:52:39 2016 -0700
+++ b/src/cpu/aarch64/vm/globals_aarch64.hpp Fri Jun 10 17:51:24 2016 +0100
@@ -48,11 +48,7 @@
// the the vep is aligned at CodeEntryAlignment whereas c2 only aligns
// the uep and the vep doesn't get real alignment but just slops on by
// only assured that the entry instruction meets the 5 byte size requirement.
-#ifdef COMPILER2
define_pd_global(intx, CodeEntryAlignment, 64);
-#else
-define_pd_global(intx, CodeEntryAlignment, 16);
-#endif // COMPILER2
define_pd_global(intx, OptoLoopAlignment, 16);
define_pd_global(intx, InlineFrequencyCount, 100);
diff -r 2d8e12787f80 -r a778398a0cdb src/cpu/aarch64/vm/stubGenerator_aarch64.cpp
--- a/src/cpu/aarch64/vm/stubGenerator_aarch64.cpp Tue Apr 19 19:52:39 2016 -0700
+++ b/src/cpu/aarch64/vm/stubGenerator_aarch64.cpp Fri Jun 10 17:51:24 2016 +0100
@@ -687,7 +687,7 @@
// count - element count
// tmp - scratch register
//
- // Destroy no registers!
+ // Destroy no registers except rscratch1 and rscratch2
//
void gen_write_ref_array_pre_barrier(Register addr, Register count, bool dest_uninitialized) {
BarrierSet* bs = Universe::heap()->barrier_set();
@@ -696,12 +696,13 @@
case BarrierSet::G1SATBCTLogging:
// With G1, don't generate the call if we statically know that the target in uninitialized
if (!dest_uninitialized) {
- __ push(RegSet::range(r0, r29), sp); // integer registers except lr & sp
+ __ push_call_clobbered_registers();
if (count == c_rarg0) {
if (addr == c_rarg1) {
// exactly backwards!!
- __ stp(c_rarg0, c_rarg1, __ pre(sp, -2 * wordSize));
- __ ldp(c_rarg1, c_rarg0, __ post(sp, -2 * wordSize));
+ __ mov(rscratch1, c_rarg0);
+ __ mov(c_rarg0, c_rarg1);
+ __ mov(c_rarg1, rscratch1);
} else {
__ mov(c_rarg1, count);
__ mov(c_rarg0, addr);
@@ -711,7 +712,7 @@
__ mov(c_rarg1, count);
}
__ call_VM_leaf(CAST_FROM_FN_PTR(address, BarrierSet::static_write_ref_array_pre), 2);
- __ pop(RegSet::range(r0, r29), sp); // integer registers except lr & sp }
+ __ pop_call_clobbered_registers();
break;
case BarrierSet::CardTableModRef:
case BarrierSet::CardTableExtension:
@@ -742,7 +743,7 @@
case BarrierSet::G1SATBCTLogging:
{
- __ push(RegSet::range(r0, r29), sp); // integer registers except lr & sp
+ __ push_call_clobbered_registers();
// must compute element count unless barrier set interface is changed (other platforms supply count)
assert_different_registers(start, end, scratch);
__ lea(scratch, Address(end, BytesPerHeapOop));
@@ -751,7 +752,7 @@
__ mov(c_rarg0, start);
__ mov(c_rarg1, scratch);
__ call_VM_leaf(CAST_FROM_FN_PTR(address, BarrierSet::static_write_ref_array_post), 2);
- __ pop(RegSet::range(r0, r29), sp); // integer registers except lr & sp }
+ __ pop_call_clobbered_registers();
}
break;
case BarrierSet::CardTableModRef:
@@ -811,7 +812,7 @@
assert_different_registers(s, d, count, rscratch1);
Label again, large, small;
- __ align(6);
+ __ align(CodeEntryAlignment);
__ bind(start);
__ cmp(count, 8);
__ br(Assembler::LO, small);
@@ -856,7 +857,7 @@
__ ret(lr);
- __ align(6);
+ __ align(CodeEntryAlignment);
__ bind(large);
// Fill 8 registers
@@ -1007,7 +1008,8 @@
}
// rscratch2 is the byte adjustment needed to align s.
__ cbz(rscratch2, aligned);
- __ lsr(rscratch2, rscratch2, exact_log2(granularity));
+ int shift = exact_log2(granularity);
+ if (shift) __ lsr(rscratch2, rscratch2, shift);
__ sub(count, count, rscratch2);
#if 0
@@ -1386,10 +1388,10 @@
// no-overlap entry point used by generate_conjoint_long_oop_copy().
//
address generate_disjoint_oop_copy(bool aligned, address *entry,
- const char *name, bool dest_uninitialized = false) {
+ const char *name, bool dest_uninitialized) {
const bool is_oop = true;
const size_t size = UseCompressedOops ? sizeof (jint) : sizeof (jlong);
- return generate_disjoint_copy(size, aligned, is_oop, entry, name);
+ return generate_disjoint_copy(size, aligned, is_oop, entry, name, dest_uninitialized);
}
// Arguments:
@@ -1404,10 +1406,11 @@
//
address generate_conjoint_oop_copy(bool aligned,
address nooverlap_target, address *entry,
- const char *name, bool dest_uninitialized = false) {
+ const char *name, bool dest_uninitialized) {
const bool is_oop = true;
const size_t size = UseCompressedOops ? sizeof (jint) : sizeof (jlong);
- return generate_conjoint_copy(size, aligned, is_oop, nooverlap_target, entry, name);
+ return generate_conjoint_copy(size, aligned, is_oop, nooverlap_target, entry,
+ name, dest_uninitialized);
}
@@ -1514,6 +1517,8 @@
}
#endif //ASSERT
+ gen_write_ref_array_pre_barrier(to, count, dest_uninitialized);
+
// save the original count
__ mov(count_save, count);
@@ -1655,9 +1660,11 @@
bool aligned = !UseCompressedOops;
StubRoutines::_arrayof_oop_disjoint_arraycopy
- = generate_disjoint_oop_copy(aligned, &entry, "arrayof_oop_disjoint_arraycopy");
+ = generate_disjoint_oop_copy(aligned, &entry, "arrayof_oop_disjoint_arraycopy",
+ /*dest_uninitialized*/false);
StubRoutines::_arrayof_oop_arraycopy
- = generate_conjoint_oop_copy(aligned, entry, &entry_oop_arraycopy, "arrayof_oop_arraycopy");
+ = generate_conjoint_oop_copy(aligned, entry, &entry_oop_arraycopy, "arrayof_oop_arraycopy",
+ /*dest_uninitialized*/false);
// Aligned versions without pre-barriers
StubRoutines::_arrayof_oop_disjoint_arraycopy_uninit
= generate_disjoint_oop_copy(aligned, &entry, "arrayof_oop_disjoint_arraycopy_uninit",
diff -r 2d8e12787f80 -r a778398a0cdb src/cpu/aarch64/vm/templateTable_aarch64.cpp
--- a/src/cpu/aarch64/vm/templateTable_aarch64.cpp Tue Apr 19 19:52:39 2016 -0700
+++ b/src/cpu/aarch64/vm/templateTable_aarch64.cpp Fri Jun 10 17:51:24 2016 +0100
@@ -3032,7 +3032,7 @@
__ null_check(r0);
switch (state) {
case itos:
- __ ldr(r0, Address(r0, r1, Address::lsl(0)));
+ __ ldrw(r0, Address(r0, r1, Address::lsl(0)));
break;
case atos:
__ load_heap_oop(r0, Address(r0, r1, Address::lsl(0)));
@@ -3052,7 +3052,7 @@
__ ldrw(r3, Address(r2, in_bytes(constantPoolCacheOopDesc::base_offset() +
ConstantPoolCacheEntry::flags_offset())));
__ tbz(r3, ConstantPoolCacheEntry::is_volatile_shift, notVolatile);
- __ membar(MacroAssembler::LoadLoad);
+ __ membar(MacroAssembler::LoadLoad | MacroAssembler::LoadStore);
__ bind(notVolatile);
}
diff -r 2d8e12787f80 -r a778398a0cdb src/share/vm/runtime/arguments.cpp
--- a/src/share/vm/runtime/arguments.cpp Tue Apr 19 19:52:39 2016 -0700
+++ b/src/share/vm/runtime/arguments.cpp Fri Jun 10 17:51:24 2016 +0100
@@ -3473,6 +3473,14 @@
set_g1_gc_flags();
}
+ if (AssumeMP && !UseSerialGC) {
+ if (FLAG_IS_DEFAULT(ParallelGCThreads) && ParallelGCThreads == 1) {
+ warning("If the number of processors is expected to increase from one, then"
+ " you should configure the number of parallel GC threads appropriately"
+ " using -XX:ParallelGCThreads=N");
+ }
+ }
+
#ifdef SERIALGC
assert(verify_serial_gc_flags(), "SerialGC unset");
#endif // SERIALGC
diff -r 2d8e12787f80 -r a778398a0cdb src/share/vm/runtime/globals.hpp
--- a/src/share/vm/runtime/globals.hpp Tue Apr 19 19:52:39 2016 -0700
+++ b/src/share/vm/runtime/globals.hpp Fri Jun 10 17:51:24 2016 +0100
@@ -480,6 +480,9 @@
lp64_product(intx, ObjectAlignmentInBytes, 8, \
"Default object alignment in bytes, 8 is minimum") \
\
+ product(bool, AssumeMP, false, \
+ "Instruct the VM to assume multiple processors are available") \
+ \
/* UseMembar is theoretically a temp flag used for memory barrier \
* removal testing. It was supposed to be removed before FCS but has \
* been re-added (see 6401008) */ \
diff -r 2d8e12787f80 -r a778398a0cdb src/share/vm/runtime/os.hpp
--- a/src/share/vm/runtime/os.hpp Tue Apr 19 19:52:39 2016 -0700
+++ b/src/share/vm/runtime/os.hpp Fri Jun 10 17:51:24 2016 +0100
@@ -198,7 +198,7 @@
// Interface for detecting multiprocessor system
static inline bool is_MP() {
assert(_processor_count > 0, "invalid processor count");
- return _processor_count > 1;
+ return _processor_count > 1 || AssumeMP;
}
static julong available_memory();
static julong physical_memory();
More information about the distro-pkg-dev
mailing list