/hg/icedtea7-forest/hotspot: 4 new changesets
andrew at icedtea.classpath.org
andrew at icedtea.classpath.org
Wed Sep 14 16:34:30 UTC 2016
changeset dd756ec2f65d in /hg/icedtea7-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea7-forest/hotspot?cmd=changeset;node=dd756ec2f65d
author: clanger
date: Wed Nov 04 16:23:08 2015 -0800
8140244, PR3171: Port fix of JDK-8075773 to AIX
Reviewed-by: stuefe, dcubed
changeset 18f9f032164c in /hg/icedtea7-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea7-forest/hotspot?cmd=changeset;node=18f9f032164c
author: simonis
date: Wed Jan 27 15:01:46 2016 +0100
8148487: PPC64: Better byte behavior
Reviewed-by: goetz, mdoerr
changeset 71fdc2163e4e in /hg/icedtea7-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea7-forest/hotspot?cmd=changeset;node=71fdc2163e4e
author: aph
date: Mon Apr 18 07:12:53 2016 +0100
8154413: AArch64: Better byte behaviour
Summary: Add support for 8132051 on AArch64
Reviewed-by: adinn
changeset b6971126dc0f in /hg/icedtea7-forest/hotspot
details: http://icedtea.classpath.org/hg/icedtea7-forest/hotspot?cmd=changeset;node=b6971126dc0f
author: andrew
date: Tue Apr 19 21:44:54 2016 +0100
PR2938: Fix build of 8148487 backport
Summary: ztos case in templateTable_ppc_64.cpp needs to match other cases in 7
diffstat:
src/cpu/aarch64/vm/c1_LIRGenerator_aarch64.cpp | 5 +-
src/cpu/aarch64/vm/interp_masm_aarch64.cpp | 41 +++++++++++++++++-
src/cpu/aarch64/vm/interp_masm_aarch64.hpp | 3 +
src/cpu/aarch64/vm/templateTable_aarch64.cpp | 59 ++++++++++++++++++++++++-
src/cpu/ppc/vm/interp_masm_ppc_64.cpp | 40 +++++++++++++++++
src/cpu/ppc/vm/interp_masm_ppc_64.hpp | 2 +
src/cpu/ppc/vm/interpreter_ppc.cpp | 10 ++++
src/cpu/ppc/vm/templateInterpreter_ppc.cpp | 12 ++++-
src/cpu/ppc/vm/templateTable_ppc_64.cpp | 61 ++++++++++++++++++++++++-
src/os/aix/vm/perfMemory_aix.cpp | 11 +++-
10 files changed, 230 insertions(+), 14 deletions(-)
diffs (truncated from 555 to 500 lines):
diff -r b3bfbaa4c0a2 -r b6971126dc0f src/cpu/aarch64/vm/c1_LIRGenerator_aarch64.cpp
--- a/src/cpu/aarch64/vm/c1_LIRGenerator_aarch64.cpp Wed Sep 07 06:06:27 2016 +0100
+++ b/src/cpu/aarch64/vm/c1_LIRGenerator_aarch64.cpp Tue Apr 19 21:44:54 2016 +0100
@@ -339,7 +339,7 @@
length.load_item();
}
}
- if (needs_store_check) {
+ if (needs_store_check || x->check_boolean()) {
value.load_item();
} else {
value.load_for_store(x->elt_type());
@@ -388,7 +388,8 @@
// Seems to be a precise
post_barrier(LIR_OprFact::address(array_addr), value.result());
} else {
- __ move(value.result(), array_addr, null_check_info);
+ LIR_Opr result = maybe_mask_boolean(x, array.result(), value.result(), null_check_info);
+ __ move(result, array_addr, null_check_info);
}
}
diff -r b3bfbaa4c0a2 -r b6971126dc0f src/cpu/aarch64/vm/interp_masm_aarch64.cpp
--- a/src/cpu/aarch64/vm/interp_masm_aarch64.cpp Wed Sep 07 06:06:27 2016 +0100
+++ b/src/cpu/aarch64/vm/interp_masm_aarch64.cpp Tue Apr 19 21:44:54 2016 +0100
@@ -52,7 +52,43 @@
#endif
-// Implementation of InterpreterMacroAssembler
+void InterpreterMacroAssembler::narrow(Register result) {
+
+ // Get method->_constMethod->_result_type
+ ldr(rscratch1, Address(rfp, frame::interpreter_frame_method_offset * wordSize));
+ ldr(rscratch1, Address(rscratch1, methodOopDesc::const_offset()));
+ ldrb(rscratch1, Address(rscratch1, constMethodOopDesc::result_type_offset()));
+
+ Label done, notBool, notByte, notChar;
+
+ // common case first
+ cmpw(rscratch1, T_INT);
+ br(Assembler::EQ, done);
+
+ // mask integer result to narrower return type.
+ cmpw(rscratch1, T_BOOLEAN);
+ br(Assembler::NE, notBool);
+ andw(result, result, 0x1);
+ b(done);
+
+ bind(notBool);
+ cmpw(rscratch1, T_BYTE);
+ br(Assembler::NE, notByte);
+ sbfx(result, result, 0, 8);
+ b(done);
+
+ bind(notByte);
+ cmpw(rscratch1, T_CHAR);
+ br(Assembler::NE, notChar);
+ ubfx(result, result, 0, 16); // truncate upper 16 bits
+ b(done);
+
+ bind(notChar);
+ sbfx(result, result, 0, 16); // sign-extend short
+
+ // Nothing to do for T_INT
+ bind(done);
+}
#ifndef CC_INTERP
@@ -90,6 +126,7 @@
verify_oop(r0, state); break;
case ltos: ldr(r0, val_addr); break;
case btos: // fall through
+ case ztos: // fall through
case ctos: // fall through
case stos: // fall through
case itos: ldrw(r0, val_addr); break;
@@ -293,6 +330,7 @@
switch (state) {
case atos: pop_ptr(); break;
case btos:
+ case ztos:
case ctos:
case stos:
case itos: pop_i(); break;
@@ -310,6 +348,7 @@
switch (state) {
case atos: push_ptr(); break;
case btos:
+ case ztos:
case ctos:
case stos:
case itos: push_i(); break;
diff -r b3bfbaa4c0a2 -r b6971126dc0f src/cpu/aarch64/vm/interp_masm_aarch64.hpp
--- a/src/cpu/aarch64/vm/interp_masm_aarch64.hpp Wed Sep 07 06:06:27 2016 +0100
+++ b/src/cpu/aarch64/vm/interp_masm_aarch64.hpp Tue Apr 19 21:44:54 2016 +0100
@@ -244,6 +244,9 @@
void update_mdp_by_constant(Register mdp_in, int constant);
void update_mdp_for_ret(Register return_bci);
+ // narrow int return value
+ void narrow(Register result);
+
void profile_taken_branch(Register mdp, Register bumped_count);
void profile_not_taken_branch(Register mdp);
void profile_call(Register mdp);
diff -r b3bfbaa4c0a2 -r b6971126dc0f src/cpu/aarch64/vm/templateTable_aarch64.cpp
--- a/src/cpu/aarch64/vm/templateTable_aarch64.cpp Wed Sep 07 06:06:27 2016 +0100
+++ b/src/cpu/aarch64/vm/templateTable_aarch64.cpp Tue Apr 19 21:44:54 2016 +0100
@@ -232,6 +232,7 @@
switch (bc) {
case Bytecodes::_fast_aputfield:
case Bytecodes::_fast_bputfield:
+ case Bytecodes::_fast_zputfield:
case Bytecodes::_fast_cputfield:
case Bytecodes::_fast_dputfield:
case Bytecodes::_fast_fputfield:
@@ -1108,6 +1109,18 @@
// r1: index
// r3: array
index_check(r3, r1); // prefer index in r1
+
+ // Need to check whether array is boolean or byte
+ // since both types share the bastore bytecode.
+ __ load_klass(r2, r3);
+ __ ldrw(r2, Address(r2, Klass::layout_helper_offset()));
+ int diffbit = Klass::layout_helper_boolean_diffbit();
+ __ andw(rscratch1, r2, diffbit);
+ Label L_skip;
+ __ cbzw(rscratch1, L_skip);
+ __ andw(r0, r0, 1); // if it is a T_BOOLEAN array, mask the stored value to 0/1
+ __ bind(L_skip);
+
__ lea(rscratch1, Address(r3, r1, Address::uxtw(0)));
__ strb(r0, Address(rscratch1,
arrayOopDesc::base_offset_in_bytes(T_BYTE)));
@@ -2204,6 +2217,13 @@
if (_desc->bytecode() == Bytecodes::_return)
__ membar(MacroAssembler::StoreStore);
+ // Narrow result if state is itos but result type is smaller.
+ // Need to narrow in the return bytecode rather than in generate_return_entry
+ // since compiled code callers expect the result to already be narrowed.
+ if (state == itos) {
+ __ narrow(r0);
+ }
+
__ remove_activation(state);
__ ret(lr);
}
@@ -2443,7 +2463,7 @@
const Address field(obj, off);
- Label Done, notByte, notInt, notShort, notChar,
+ Label Done, notByte, notBool, notInt, notShort, notChar,
notLong, notFloat, notObj, notDouble;
// x86 uses a shift and mask or wings it with a shift plus assert
@@ -2463,6 +2483,20 @@
__ b(Done);
__ bind(notByte);
+ __ cmp(flags, ztos);
+ __ br(Assembler::NE, notBool);
+
+ // ztos (same code as btos)
+ __ ldrsb(r0, field);
+ __ push(ztos);
+ // Rewrite bytecode to be faster
+ if (!is_static) {
+ // use btos rewriting, no truncating to t/f bit is needed for getfield.
+ patch_bytecode(Bytecodes::_fast_bgetfield, bc, r1);
+ }
+ __ b(Done);
+
+ __ bind(notBool);
__ cmp(flags, atos);
__ br(Assembler::NE, notObj);
// atos
@@ -2655,7 +2689,7 @@
// field address
const Address field(obj, off);
- Label notByte, notInt, notShort, notChar,
+ Label notByte, notBool, notInt, notShort, notChar,
notLong, notFloat, notObj, notDouble;
// x86 uses a shift and mask or wings it with a shift plus assert
@@ -2677,6 +2711,22 @@
}
__ bind(notByte);
+ __ cmp(flags, ztos);
+ __ br(Assembler::NE, notBool);
+
+ // ztos
+ {
+ __ pop(ztos);
+ if (!is_static) pop_and_check_object(obj);
+ __ andw(r0, r0, 0x1);
+ __ strb(r0, field);
+ if (!is_static) {
+ patch_bytecode(Bytecodes::_fast_zputfield, bc, r1, true, byte_no);
+ }
+ __ b(Done);
+ }
+
+ __ bind(notBool);
__ cmp(flags, atos);
__ br(Assembler::NE, notObj);
@@ -2827,6 +2877,7 @@
switch (bytecode()) { // load values into the jvalue object
case Bytecodes::_fast_aputfield: __ push_ptr(r0); break;
case Bytecodes::_fast_bputfield: // fall through
+ case Bytecodes::_fast_zputfield: // fall through
case Bytecodes::_fast_sputfield: // fall through
case Bytecodes::_fast_cputfield: // fall through
case Bytecodes::_fast_iputfield: __ push_i(r0); break;
@@ -2852,6 +2903,7 @@
switch (bytecode()) { // restore tos values
case Bytecodes::_fast_aputfield: __ pop_ptr(r0); break;
case Bytecodes::_fast_bputfield: // fall through
+ case Bytecodes::_fast_zputfield: // fall through
case Bytecodes::_fast_sputfield: // fall through
case Bytecodes::_fast_cputfield: // fall through
case Bytecodes::_fast_iputfield: __ pop_i(r0); break;
@@ -2907,6 +2959,9 @@
case Bytecodes::_fast_iputfield:
__ strw(r0, field);
break;
+ case Bytecodes::_fast_zputfield:
+ __ andw(r0, r0, 0x1); // boolean is true if LSB is 1
+ // fall through to bputfield
case Bytecodes::_fast_bputfield:
__ strb(r0, field);
break;
diff -r b3bfbaa4c0a2 -r b6971126dc0f src/cpu/ppc/vm/interp_masm_ppc_64.cpp
--- a/src/cpu/ppc/vm/interp_masm_ppc_64.cpp Wed Sep 07 06:06:27 2016 +0100
+++ b/src/cpu/ppc/vm/interp_masm_ppc_64.cpp Tue Apr 19 21:44:54 2016 +0100
@@ -175,6 +175,7 @@
case ltos: ld(R17_tos, in_bytes(JvmtiThreadState::earlyret_value_offset()), RjvmtiState);
break;
case btos: // fall through
+ case ztos: // fall through
case ctos: // fall through
case stos: // fall through
case itos: lwz(R17_tos, in_bytes(JvmtiThreadState::earlyret_value_offset()), RjvmtiState);
@@ -303,6 +304,7 @@
switch (state) {
case atos: push_ptr(); break;
case btos:
+ case ztos:
case ctos:
case stos:
case itos: push_i(); break;
@@ -318,6 +320,7 @@
switch (state) {
case atos: pop_ptr(); break;
case btos:
+ case ztos:
case ctos:
case stos:
case itos: pop_i(); break;
@@ -750,6 +753,43 @@
stdux(Rscratch2, R1_SP, Rscratch1); // atomically set *(SP = top_frame_sp) = **SP
}
+void InterpreterMacroAssembler::narrow(Register result) {
+ Register ret_type = R11_scratch1;
+ ld(R11_scratch1, in_bytes(methodOopDesc::const_offset()), R19_method);
+ lbz(ret_type, in_bytes(constMethodOopDesc::result_type_offset()), R11_scratch1);
+
+ Label notBool, notByte, notChar, done;
+
+ // common case first
+ cmpwi(CCR0, ret_type, T_INT);
+ beq(CCR0, done);
+
+ cmpwi(CCR0, ret_type, T_BOOLEAN);
+ bne(CCR0, notBool);
+ andi(result, result, 0x1);
+ b(done);
+
+ bind(notBool);
+ cmpwi(CCR0, ret_type, T_BYTE);
+ bne(CCR0, notByte);
+ extsb(result, result);
+ b(done);
+
+ bind(notByte);
+ cmpwi(CCR0, ret_type, T_CHAR);
+ bne(CCR0, notChar);
+ andi(result, result, 0xffff);
+ b(done);
+
+ bind(notChar);
+ // cmpwi(CCR0, ret_type, T_SHORT); // all that's left
+ // bne(CCR0, done);
+ extsh(result, result);
+
+ // Nothing to do for T_INT
+ bind(done);
+}
+
// Remove activation.
//
// Unlock the receiver if this is a synchronized method.
diff -r b3bfbaa4c0a2 -r b6971126dc0f src/cpu/ppc/vm/interp_masm_ppc_64.hpp
--- a/src/cpu/ppc/vm/interp_masm_ppc_64.hpp Wed Sep 07 06:06:27 2016 +0100
+++ b/src/cpu/ppc/vm/interp_masm_ppc_64.hpp Tue Apr 19 21:44:54 2016 +0100
@@ -149,6 +149,8 @@
void get_cpool_and_tags(Register Rcpool, Register Rtags);
void is_a(Label& L);
+ void narrow(Register result);
+
// Java Call Helpers
void call_from_interpreter(Register Rtarget_method, Register Rret_addr, Register Rscratch1, Register Rscratch2);
diff -r b3bfbaa4c0a2 -r b6971126dc0f src/cpu/ppc/vm/interpreter_ppc.cpp
--- a/src/cpu/ppc/vm/interpreter_ppc.cpp Wed Sep 07 06:06:27 2016 +0100
+++ b/src/cpu/ppc/vm/interpreter_ppc.cpp Tue Apr 19 21:44:54 2016 +0100
@@ -656,6 +656,16 @@
__ blr();
}
+ if (branch_table[ztos] == 0) { // generate only once
+ __ align(32, 28, 28); // align load
+ __ fence(); // volatile entry point (one instruction before non-volatile_entry point)
+ branch_table[ztos] = __ pc(); // non-volatile_entry point
+ __ lbzx(R3_RET, Rclass_or_obj, Roffset);
+ __ extsb(R3_RET, R3_RET);
+ __ beq(CCR6, Lacquire);
+ __ blr();
+ }
+
if (branch_table[ctos] == 0) { // generate only once
__ align(32, 28, 28); // align load
__ fence(); // volatile entry point (one instruction before non-volatile_entry point)
diff -r b3bfbaa4c0a2 -r b6971126dc0f src/cpu/ppc/vm/templateInterpreter_ppc.cpp
--- a/src/cpu/ppc/vm/templateInterpreter_ppc.cpp Wed Sep 07 06:06:27 2016 +0100
+++ b/src/cpu/ppc/vm/templateInterpreter_ppc.cpp Tue Apr 19 21:44:54 2016 +0100
@@ -157,6 +157,7 @@
switch (state) {
case ltos:
case btos:
+ case ztos:
case ctos:
case stos:
case atos:
@@ -216,6 +217,7 @@
switch (state) {
case ltos:
case btos:
+ case ztos:
case ctos:
case stos:
case atos:
@@ -1661,12 +1663,14 @@
// Copied from TemplateTable::_return.
// Restoration of lr done by remove_activation.
switch (state) {
+ // Narrow result if state is itos but result type is smaller.
+ case itos: __ narrow(R17_tos); /* fall through */
case ltos:
case btos:
+ case ztos:
case ctos:
case stos:
- case atos:
- case itos: __ mr(R3_RET, R17_tos); break;
+ case atos: __ mr(R3_RET, R17_tos); break;
case ftos:
case dtos: __ fmr(F1_RET, F15_ftos); break;
case vtos: // This might be a constructor. Final fields (and volatile fields on PPC64) need
@@ -1736,6 +1740,10 @@
bname = "trace_code_btos {";
tsize = 2;
break;
+ case ztos:
+ bname = "trace_code_ztos {";
+ tsize = 2;
+ break;
case ctos:
bname = "trace_code_ctos {";
tsize = 2;
diff -r b3bfbaa4c0a2 -r b6971126dc0f src/cpu/ppc/vm/templateTable_ppc_64.cpp
--- a/src/cpu/ppc/vm/templateTable_ppc_64.cpp Wed Sep 07 06:06:27 2016 +0100
+++ b/src/cpu/ppc/vm/templateTable_ppc_64.cpp Tue Apr 19 21:44:54 2016 +0100
@@ -175,6 +175,7 @@
switch (new_bc) {
case Bytecodes::_fast_aputfield:
case Bytecodes::_fast_bputfield:
+ case Bytecodes::_fast_zputfield:
case Bytecodes::_fast_cputfield:
case Bytecodes::_fast_dputfield:
case Bytecodes::_fast_fputfield:
@@ -1011,9 +1012,21 @@
Rarray = R12_scratch2,
Rscratch = R3_ARG1;
__ pop_i(Rindex);
+ __ pop_ptr(Rarray);
// tos: val
- // Rarray: array ptr (popped by index_check)
- __ index_check(Rarray, Rindex, 0, Rscratch, Rarray);
+
+ // Need to check whether array is boolean or byte
+ // since both types share the bastore bytecode.
+ __ load_klass(Rscratch, Rarray);
+ __ lwz(Rscratch, in_bytes(Klass::layout_helper_offset()), Rscratch);
+ int diffbit = exact_log2(Klass::layout_helper_boolean_diffbit());
+ __ testbitdi(CCR0, R0, Rscratch, diffbit);
+ Label L_skip;
+ __ bfalse(CCR0, L_skip);
+ __ andi(R17_tos, R17_tos, 1); // if it is a T_BOOLEAN array, mask the stored value to 0/1
+ __ bind(L_skip);
+
+ __ index_check_without_pop(Rarray, Rindex, 0, Rscratch, Rarray);
__ stb(R17_tos, arrayOopDesc::base_offset_in_bytes(T_BYTE), Rarray);
}
@@ -2143,12 +2156,16 @@
__ remove_activation(state, /* throw_monitor_exception */ true);
// Restoration of lr done by remove_activation.
switch (state) {
+ // Narrow result if state is itos but result type is smaller.
+ // Need to narrow in the return bytecode rather than in generate_return_entry
+ // since compiled code callers expect the result to already be narrowed.
+ case itos: __ narrow(R17_tos); /* fall through */
case ltos:
case btos:
+ case ztos:
case ctos:
case stos:
- case atos:
- case itos: __ mr(R3_RET, R17_tos); break;
+ case atos: __ mr(R3_RET, R17_tos); break;
case ftos:
case dtos: __ fmr(F1_RET, F15_ftos); break;
case vtos: // This might be a constructor. Final fields (and volatile fields on PPC64) need
@@ -2567,6 +2584,21 @@
__ b(Lacquire);
__ align(32, 28, 28); // Align load.
+ // __ bind(Lztos); (same code as btos)
+ __ fence(); // Volatile entry point (one instruction before non-volatile_entry point).
+ assert(branch_table[ztos] == 0, "can't compute twice");
+ branch_table[ztos] = __ pc(); // non-volatile_entry point
+ __ lbzx(R17_tos, Rclass_or_obj, Roffset);
+ __ extsb(R17_tos, R17_tos);
+ __ push(ztos);
+ if (!is_static) {
+ // use btos rewriting, no truncating to t/f bit is needed for getfield.
+ patch_bytecode(Bytecodes::_fast_bgetfield, Rbc, Rscratch);
+ }
+ __ beq(CCR6, Lacquire); // Volatile?
+ __ dispatch_epilog(vtos, Bytecodes::length_for(bytecode()));
+
+ __ align(32, 28, 28); // Align load.
// __ bind(Lctos);
__ fence(); // Volatile entry point (one instruction before non-volatile_entry point).
assert(branch_table[ctos] == 0, "can't compute twice");
@@ -2658,6 +2690,7 @@
case Bytecodes::_fast_aputfield: __ push_ptr(); offs += Interpreter::stackElementSize; break;
case Bytecodes::_fast_iputfield: // Fall through
case Bytecodes::_fast_bputfield: // Fall through
+ case Bytecodes::_fast_zputfield: // Fall through
case Bytecodes::_fast_cputfield: // Fall through
case Bytecodes::_fast_sputfield: __ push_i(); offs += Interpreter::stackElementSize; break;
case Bytecodes::_fast_lputfield: __ push_l(); offs += 2*Interpreter::stackElementSize; break;
@@ -2698,6 +2731,7 @@
case Bytecodes::_fast_aputfield: __ pop_ptr(); break;
case Bytecodes::_fast_iputfield: // Fall through
case Bytecodes::_fast_bputfield: // Fall through
+ case Bytecodes::_fast_zputfield: // Fall through
case Bytecodes::_fast_cputfield: // Fall through
case Bytecodes::_fast_sputfield: __ pop_i(); break;
case Bytecodes::_fast_lputfield: __ pop_l(); break;
@@ -2854,6 +2888,22 @@
__ b(Lexit);
__ align(32, 28, 28); // Align pop.
+ // __ bind(Lztos);
+ __ release(); // Volatile entry point (one instruction before non-volatile_entry point).
+ assert(branch_table[ztos] == 0, "can't compute twice");
+ branch_table[ztos] = __ pc(); // non-volatile_entry point
+ __ pop(ztos);
+ if (!is_static) { pop_and_check_object(Rclass_or_obj); } // Kills R11_scratch1.
+ __ andi(R17_tos, R17_tos, 0x1);
+ __ stbx(R17_tos, Rclass_or_obj, Roffset);
+ if (!is_static) { patch_bytecode(Bytecodes::_fast_zputfield, Rbc, Rscratch, true, byte_no); }
+ if (!support_IRIW_for_not_multiple_copy_atomic_cpu) {
+ __ bne(CR_is_vol, Lexit); // Non-volatile?
+ __ fence();
+ }
+ __ b(Lexit);
+
More information about the distro-pkg-dev
mailing list