/hg/icedtea8-forest/hotspot: 3 new changesets
andrew at icedtea.classpath.org
andrew at icedtea.classpath.org
Fri Jan 13 17:03:42 UTC 2017
changeset 7fc218dce059 in /hg/icedtea8-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea8-forest/hotspot?cmd=changeset;node=7fc218dce059
author: thartmann
date: Thu Jan 12 18:58:53 2017 +0000
8130309, PR3280: Need to bailout cleanly if creation of stubs fails when codecache is out of space
Summary: Check for failed expansion of stub section in code buffer and bailout.
Reviewed-by: kvn, adinn, dlong, roland, twisti
changeset 510bf6178ea4 in /hg/icedtea8-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea8-forest/hotspot?cmd=changeset;node=510bf6178ea4
author: adinn
date: Thu Jan 12 19:02:07 2017 +0000
8132875, PR3280: AArch64: Fix error introduced into AArch64 CodeCache by commit for 8130309
Summary: The fix for issue 8130309 introduced several errors into the AArch64 codecache routines
Reviewed-by: aph, thartmann, kvn
changeset f73f57792257 in /hg/icedtea8-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea8-forest/hotspot?cmd=changeset;node=f73f57792257
author: andrew
date: Fri Jan 13 17:09:34 2017 +0000
Added tag icedtea-3.3.0pre01 for changeset 510bf6178ea4
diffstat:
.hgtags | 1 +
src/cpu/aarch64/vm/aarch64.ad | 62 +++++++++++++++++++------
src/cpu/aarch64/vm/c1_CodeStubs_aarch64.cpp | 9 +++-
src/cpu/aarch64/vm/c1_LIRAssembler_aarch64.cpp | 12 ++++-
src/cpu/aarch64/vm/c1_LIRAssembler_aarch64.hpp | 3 +
src/cpu/aarch64/vm/compiledIC_aarch64.cpp | 10 ++-
src/cpu/aarch64/vm/macroAssembler_aarch64.cpp | 21 +++++---
src/cpu/aarch64/vm/macroAssembler_aarch64.hpp | 6 +-
src/share/vm/code/compiledIC.hpp | 2 +-
9 files changed, 91 insertions(+), 35 deletions(-)
diffs (342 lines):
diff -r 4bc7ad288df4 -r f73f57792257 .hgtags
--- a/.hgtags Thu Jan 12 06:59:37 2017 +0000
+++ b/.hgtags Fri Jan 13 17:09:34 2017 +0000
@@ -962,3 +962,4 @@
7bc1061f52cfc5ce4cbfd42dd2dcdc91e7efce72 jdk8u112-b13
c2c4db2a42a215c98a4f027edb8bbb00dd62d9b9 jdk8u112-b14
b28d012a24cab8f4ceeee0c9d3252969757423ed jdk8u112-b15
+510bf6178ea48859804d69715a78b82b8d2c58d3 icedtea-3.3.0pre01
diff -r 4bc7ad288df4 -r f73f57792257 src/cpu/aarch64/vm/aarch64.ad
--- a/src/cpu/aarch64/vm/aarch64.ad Thu Jan 12 06:59:37 2017 +0000
+++ b/src/cpu/aarch64/vm/aarch64.ad Fri Jan 13 17:09:34 2017 +0000
@@ -3349,9 +3349,11 @@
// Note that the code buffer's insts_mark is always relative to insts.
// That's why we must use the macroassembler to generate a handler.
MacroAssembler _masm(&cbuf);
- address base =
- __ start_a_stub(size_exception_handler());
- if (base == NULL) return 0; // CodeBuffer::expand failed
+ address base = __ start_a_stub(size_exception_handler());
+ if (base == NULL) {
+ ciEnv::current()->record_failure("CodeCache is full");
+ return 0; // CodeBuffer::expand failed
+ }
int offset = __ offset();
__ far_jump(RuntimeAddress(OptoRuntime::exception_blob()->entry_point()));
assert(__ offset() - offset <= (int) size_exception_handler(), "overflow");
@@ -3365,9 +3367,11 @@
// Note that the code buffer's insts_mark is always relative to insts.
// That's why we must use the macroassembler to generate a handler.
MacroAssembler _masm(&cbuf);
- address base =
- __ start_a_stub(size_deopt_handler());
- if (base == NULL) return 0; // CodeBuffer::expand failed
+ address base = __ start_a_stub(size_deopt_handler());
+ if (base == NULL) {
+ ciEnv::current()->record_failure("CodeCache is full");
+ return 0; // CodeBuffer::expand failed
+ }
int offset = __ offset();
__ adr(lr, __ pc());
@@ -4643,18 +4647,27 @@
address mark = __ pc();
address addr = (address)$meth$$method;
+ address call;
if (!_method) {
// A call to a runtime wrapper, e.g. new, new_typeArray_Java, uncommon_trap.
- __ trampoline_call(Address(addr, relocInfo::runtime_call_type), &cbuf);
+ call = __ trampoline_call(Address(addr, relocInfo::runtime_call_type), &cbuf);
} else if (_optimized_virtual) {
- __ trampoline_call(Address(addr, relocInfo::opt_virtual_call_type), &cbuf);
+ call = __ trampoline_call(Address(addr, relocInfo::opt_virtual_call_type), &cbuf);
} else {
- __ trampoline_call(Address(addr, relocInfo::static_call_type), &cbuf);
+ call = __ trampoline_call(Address(addr, relocInfo::static_call_type), &cbuf);
+ }
+ if (call == NULL) {
+ ciEnv::current()->record_failure("CodeCache is full");
+ return;
}
if (_method) {
// Emit stub for static call
- CompiledStaticCall::emit_to_interp_stub(cbuf, mark);
+ address stub = CompiledStaticCall::emit_to_interp_stub(cbuf, mark);
+ if (stub == NULL) {
+ ciEnv::current()->record_failure("CodeCache is full");
+ return;
+ }
}
%}
@@ -4668,18 +4681,27 @@
address mark = __ pc();
address addr = (address)$meth$$method;
+ address call;
if (!_method) {
// A call to a runtime wrapper, e.g. new, new_typeArray_Java, uncommon_trap.
- __ trampoline_call(Address(addr, relocInfo::runtime_call_type), &cbuf);
+ call = __ trampoline_call(Address(addr, relocInfo::runtime_call_type), &cbuf);
} else if (_optimized_virtual) {
- __ trampoline_call(Address(addr, relocInfo::opt_virtual_call_type), &cbuf);
+ call = __ trampoline_call(Address(addr, relocInfo::opt_virtual_call_type), &cbuf);
} else {
- __ trampoline_call(Address(addr, relocInfo::static_call_type), &cbuf);
+ call = __ trampoline_call(Address(addr, relocInfo::static_call_type), &cbuf);
+ }
+ if (call == NULL) {
+ ciEnv::current()->record_failure("CodeCache is full");
+ return;
}
if (_method) {
// Emit stub for static call
- CompiledStaticCall::emit_to_interp_stub(cbuf, mark);
+ address stub = CompiledStaticCall::emit_to_interp_stub(cbuf, mark);
+ if (stub == NULL) {
+ ciEnv::current()->record_failure("CodeCache is full");
+ return;
+ }
}
// now restore sp
@@ -4688,7 +4710,11 @@
enc_class aarch64_enc_java_dynamic_call(method meth) %{
MacroAssembler _masm(&cbuf);
- __ ic_call((address)$meth$$method);
+ address call = __ ic_call((address)$meth$$method);
+ if (call == NULL) {
+ ciEnv::current()->record_failure("CodeCache is full");
+ return;
+ }
%}
enc_class aarch64_enc_call_epilog() %{
@@ -4709,7 +4735,11 @@
address entry = (address)$meth$$method;
CodeBlob *cb = CodeCache::find_blob(entry);
if (cb) {
- __ trampoline_call(Address(entry, relocInfo::runtime_call_type));
+ address call = __ trampoline_call(Address(entry, relocInfo::runtime_call_type));
+ if (call == NULL) {
+ ciEnv::current()->record_failure("CodeCache is full");
+ return;
+ }
} else {
int gpcnt;
int fpcnt;
diff -r 4bc7ad288df4 -r f73f57792257 src/cpu/aarch64/vm/c1_CodeStubs_aarch64.cpp
--- a/src/cpu/aarch64/vm/c1_CodeStubs_aarch64.cpp Thu Jan 12 06:59:37 2017 +0000
+++ b/src/cpu/aarch64/vm/c1_CodeStubs_aarch64.cpp Fri Jan 13 17:09:34 2017 +0000
@@ -399,9 +399,16 @@
ce->align_call(lir_static_call);
ce->emit_static_call_stub();
+ if (ce->compilation()->bailed_out()) {
+ return; // CodeCache is full
+ }
Address resolve(SharedRuntime::get_resolve_static_call_stub(),
relocInfo::static_call_type);
- __ trampoline_call(resolve);
+ address call = __ trampoline_call(resolve);
+ if (call == NULL) {
+ ce->bailout("trampoline stub overflow");
+ return;
+ }
ce->add_call_info_here(info());
#ifndef PRODUCT
diff -r 4bc7ad288df4 -r f73f57792257 src/cpu/aarch64/vm/c1_LIRAssembler_aarch64.cpp
--- a/src/cpu/aarch64/vm/c1_LIRAssembler_aarch64.cpp Thu Jan 12 06:59:37 2017 +0000
+++ b/src/cpu/aarch64/vm/c1_LIRAssembler_aarch64.cpp Fri Jan 13 17:09:34 2017 +0000
@@ -2058,13 +2058,21 @@
void LIR_Assembler::call(LIR_OpJavaCall* op, relocInfo::relocType rtype) {
- __ trampoline_call(Address(op->addr(), rtype));
+ address call = __ trampoline_call(Address(op->addr(), rtype));
+ if (call == NULL) {
+ bailout("trampoline stub overflow");
+ return;
+ }
add_call_info(code_offset(), op->info());
}
void LIR_Assembler::ic_call(LIR_OpJavaCall* op) {
- __ ic_call(op->addr());
+ address call = __ ic_call(op->addr());
+ if (call == NULL) {
+ bailout("trampoline stub overflow");
+ return;
+ }
add_call_info(code_offset(), op->info());
}
diff -r 4bc7ad288df4 -r f73f57792257 src/cpu/aarch64/vm/c1_LIRAssembler_aarch64.hpp
--- a/src/cpu/aarch64/vm/c1_LIRAssembler_aarch64.hpp Thu Jan 12 06:59:37 2017 +0000
+++ b/src/cpu/aarch64/vm/c1_LIRAssembler_aarch64.hpp Fri Jan 13 17:09:34 2017 +0000
@@ -27,6 +27,9 @@
#ifndef CPU_X86_VM_C1_LIRASSEMBLER_X86_HPP
#define CPU_X86_VM_C1_LIRASSEMBLER_X86_HPP
+// ArrayCopyStub needs access to bailout
+friend class ArrayCopyStub;
+
private:
int array_element_size(BasicType type) const;
diff -r 4bc7ad288df4 -r f73f57792257 src/cpu/aarch64/vm/compiledIC_aarch64.cpp
--- a/src/cpu/aarch64/vm/compiledIC_aarch64.cpp Thu Jan 12 06:59:37 2017 +0000
+++ b/src/cpu/aarch64/vm/compiledIC_aarch64.cpp Fri Jan 13 17:09:34 2017 +0000
@@ -50,7 +50,7 @@
// ----------------------------------------------------------------------------
#define __ _masm.
-void CompiledStaticCall::emit_to_interp_stub(CodeBuffer &cbuf, address mark) {
+address CompiledStaticCall::emit_to_interp_stub(CodeBuffer &cbuf, address mark) {
// Stub is fixed up when the corresponding call is converted from
// calling compiled code to calling interpreted code.
// movq rmethod, 0
@@ -62,10 +62,11 @@
// That's why we must use the macroassembler to generate a stub.
MacroAssembler _masm(&cbuf);
- address base = __ start_a_stub(to_interp_stub_size()*2);
-
+ address base = __ start_a_stub(to_interp_stub_size());
int offset = __ offset();
- if (base == NULL) return; // CodeBuffer::expand failed
+ if (base == NULL) {
+ return NULL; // CodeBuffer::expand failed
+ }
// static stub relocation stores the instruction address of the call
__ relocate(static_stub_Relocation::spec(mark));
// static stub relocation also tags the Method* in the code-stream.
@@ -75,6 +76,7 @@
assert((__ offset() - offset) <= (int)to_interp_stub_size(), "stub too big");
__ end_a_stub();
+ return base;
}
#undef __
diff -r 4bc7ad288df4 -r f73f57792257 src/cpu/aarch64/vm/macroAssembler_aarch64.cpp
--- a/src/cpu/aarch64/vm/macroAssembler_aarch64.cpp Thu Jan 12 06:59:37 2017 +0000
+++ b/src/cpu/aarch64/vm/macroAssembler_aarch64.cpp Fri Jan 13 17:09:34 2017 +0000
@@ -676,7 +676,7 @@
// Maybe emit a call via a trampoline. If the code cache is small
// trampolines won't be emitted.
-void MacroAssembler::trampoline_call(Address entry, CodeBuffer *cbuf) {
+address MacroAssembler::trampoline_call(Address entry, CodeBuffer *cbuf) {
assert(entry.rspec().type() == relocInfo::runtime_call_type
|| entry.rspec().type() == relocInfo::opt_virtual_call_type
|| entry.rspec().type() == relocInfo::static_call_type
@@ -685,7 +685,10 @@
unsigned int start_offset = offset();
#ifdef COMPILER2
if (far_branches() && !Compile::current()->in_scratch_emit_size()) {
- emit_trampoline_stub(offset(), entry.target());
+ address stub = emit_trampoline_stub(start_offset, entry.target());
+ if (stub == NULL) {
+ return NULL; // CodeCache is full
+ }
}
#endif
@@ -700,6 +703,8 @@
#else
bl(entry.target());
#endif
+ // just need to return a non-null address
+ return pc();
}
@@ -714,14 +719,12 @@
// load the call target from the constant pool
// branch (LR still points to the call site above)
-void MacroAssembler::emit_trampoline_stub(int insts_call_instruction_offset,
+address MacroAssembler::emit_trampoline_stub(int insts_call_instruction_offset,
address dest) {
#ifdef COMPILER2
address stub = start_a_stub(Compile::MAX_stubs_size/2);
if (stub == NULL) {
- start_a_stub(Compile::MAX_stubs_size/2);
- Compile::current()->env()->record_out_of_memory_failure();
- return;
+ return NULL; // CodeBuffer::expand failed
}
// Create a trampoline stub relocation which relates this trampoline stub
@@ -748,18 +751,20 @@
assert(is_NativeCallTrampolineStub_at(stub_start_addr), "doesn't look like a trampoline");
end_a_stub();
+ return stub;
#else
ShouldNotReachHere();
+ return NULL;
#endif
}
-void MacroAssembler::ic_call(address entry) {
+address MacroAssembler::ic_call(address entry) {
RelocationHolder rh = virtual_call_Relocation::spec(pc());
// address const_ptr = long_constant((jlong)Universe::non_oop_word());
// unsigned long offset;
// ldr_constant(rscratch2, const_ptr);
movptr(rscratch2, (uintptr_t)Universe::non_oop_word());
- trampoline_call(Address(entry, rh));
+ return trampoline_call(Address(entry, rh));
}
// Implementation of call_VM versions
diff -r 4bc7ad288df4 -r f73f57792257 src/cpu/aarch64/vm/macroAssembler_aarch64.hpp
--- a/src/cpu/aarch64/vm/macroAssembler_aarch64.hpp Thu Jan 12 06:59:37 2017 +0000
+++ b/src/cpu/aarch64/vm/macroAssembler_aarch64.hpp Fri Jan 13 17:09:34 2017 +0000
@@ -591,7 +591,7 @@
static int patch_oop(address insn_addr, address o);
- void emit_trampoline_stub(int insts_call_instruction_offset, address target);
+ address emit_trampoline_stub(int insts_call_instruction_offset, address target);
// The following 4 methods return the offset of the appropriate move instruction
@@ -998,7 +998,7 @@
// Calls
- void trampoline_call(Address entry, CodeBuffer *cbuf = NULL);
+ address trampoline_call(Address entry, CodeBuffer *cbuf = NULL);
static bool far_branches() {
return ReservedCodeCacheSize > branch_range;
@@ -1018,7 +1018,7 @@
}
// Emit the CompiledIC call idiom
- void ic_call(address entry);
+ address ic_call(address entry);
public:
diff -r 4bc7ad288df4 -r f73f57792257 src/share/vm/code/compiledIC.hpp
--- a/src/share/vm/code/compiledIC.hpp Thu Jan 12 06:59:37 2017 +0000
+++ b/src/share/vm/code/compiledIC.hpp Fri Jan 13 17:09:34 2017 +0000
@@ -324,7 +324,7 @@
// Code
#if defined AARCH64 && !defined ZERO
- static void emit_to_interp_stub(CodeBuffer &cbuf, address mark);
+ static address emit_to_interp_stub(CodeBuffer &cbuf, address mark);
#else
static address emit_to_interp_stub(CodeBuffer &cbuf);
#endif
More information about the distro-pkg-dev
mailing list