8234160: ZGC: Enable optimized mitigation for Intel jcc erratum in C2 load barrier
Vladimir Ivanov
vladimir.x.ivanov at oracle.com
Wed Feb 19 15:00:23 UTC 2020
Hi Erik,
> I have talked to Vladimir, and I think we at this point in time agree
> about doing this.
Yes, we agreed that it's reasonable to take Erik's implementation as the
stop-the-gap fix and come up with a more comprehensive fix later. Intel
folks have been experimenting with MacroAssembler-based solution and the
results are promising. But it needs more time.
> As a reminder, here is my webrev that I proposed (+ rebase that I just
> made):
> http://cr.openjdk.java.net/~eosterlund/8234160/webrev.01/
Can x86-specific changes in output.cpp & node.hpp be moved to a
arch-specific location? x86.ad maybe?
Have you considered extending MachNode::compute_padding() to do the job?
Best regards,
Vladimir Ivanov
> On 11/25/19 4:31 PM, Vladimir Ivanov wrote:
>> Hi Erik,
>>
>>>> But I'd include stubs as well. Many of them are extensively used
>>>> from C2-generated code.
>>>
>>> Okay. Any specific stubs you have in mind?If there are some critical
>>> ones, we can sprinkle some scope objects like I did in the ZGC code.
>>
>> There are intrinsics for compressed strings [1], numerous copy stubs
>> [2], trigonometric functions [3].
>>
>> It would be unfortunate if we have to go over all that code and
>> manually instrument all the places where problematic instructions are
>> issued. Moreover, the process has to be repeated for new code being
>> added over time.
>>
>>> I do have concerns though about injecting magic into the
>>> MacroAssembler that tries to solve this automagically on the assembly
>>> level, by having the assembler spit out different
>>> instructions than you requested.
>>> The following comment from assembler.hpp captures my thought exactly:
>>>
>>> 207: // The Abstract Assembler: Pure assembler doing NO optimizations
>>> on the
>>> 208: // instruction level; i.e., what you write is what you get.
>>> 209: // The Assembler is generating code into a CodeBuffer.
>>
>> While I see that Assembler follows that (instruction per method),
>> MacroAssembler does not: there are cases when generated code differ
>> depending on runtime flags (e.g., verification code) or input values
>> (e.g., whether AddressLiteral is reachable or not).
>>
>>> I think it is desirable to keep the property that when we tell the
>>> *Assembler to generate a __ cmp(); __ jcc(); it will do exactly that.
>>> When such assumptions break, any code that has calculated the size of
>>> instructions, making assumptions about their size, will fail.
>>> For example, any MachNode with hardcoded size() might underestimate
>>> how much memory is really needed, and code such as nmethod entry
>>> barriers
>>> that have calculated the offset to the cmp immediate might suddenly
>>> stop working because. There is similar code for oop maps where we
>>> calculate offsets into mach nodes with oop maps to describe the PC
>>> after a call, which will stop working:
>>>
>>> // !!!!! Special hack to get all types of calls to specify the byte
>>> offset
>>> // from the start of the call to the point where the return
>>> address
>>> // will point.
>>> int MachCallStaticJavaNode::ret_addr_offset()
>>> {
>>> int offset = 5; // 5 bytes from start of call to where return
>>> address points
>>> offset += clear_avx_size();
>>> return offset;
>>> }
>>>
>>> Basically, I think you might be able to mitigate more branches on the
>>> MacroAssembler layer, but I think it would also be more risky, as
>>> code that was
>>> not built for having random size will start failing, in places we
>>> didn't think of.I can think of a few, and feel like there are
>>> probably other places I have not thought about.
>>>
>>> So from that point of view, I think I would rather to this on Mach
>>> nodes where it is safe, and I think we can catch the most important
>>> ones there,
>>> and miss a few branches that the macro assembler would have found
>>> with magic, than apply it to all branches and hope we find all the
>>> bugs due to unexpected magic.
>>>
>>> Do you agree? Or perhaps I misunderstood what you are suggesting.
>>
>> You raise a valid point: there are places in the VM which rely on
>> hard-coded instruction sequences. If such instruction changes, all
>> relevant places have to be adjusted. And JVM is already very cautious
>> about such cases.
>>
>> I agree with you that MacroAssembler-based more risky, but IMO the
>> risk is modest (few places are affected) and manageable (dedicated
>> stress mode should greatly improve test effectiveness).
>>
>> My opinion is that if we are satisfied with the coverage C2 CFG
>> instrumentation provides and don't expect any more work on
>> mitigations, then there's no motivation in investing into
>> MacroAssembler-based approach.
>>
>> Otherwise, there are basically 2 options:
>>
>> * "opt-in": explicitly mark all the places where mitigations are
>> applied, by default nothing is mitigated
>>
>> * "opt-out": mitigate everything unless mitigations are explicitly
>> disabled
>>
>> Both approaches provide fine-grained control over what's being
>> mitigated, but with "opt-out" there's more code to care about: it's
>> easy to miss important cases and too tempting to enable more than we
>> are 100% certain about.
>>
>> Both can be applied to individual CFG nodes and make CFG
>> instrumentation redundant.
>>
>> But if there's a need to instrument large portions of (macro)assembly
>> code, then IMO opt-in adds too much in terms of work required, noise
>> (on code level), maintenance, and burden for future code changes. So,
>> I don't consider it as a feasible option in such situation.
>>
>> It looks like a mixture of opt-in (explicitly enable in some context:
>> in C2 during code emission, particular stub generation, etc) and
>> opt-out (on the level of individual instructions) gives the best of
>> both approaches.
>>
>> But, again, if C2 CFG instrumentation is good enough, then it'll be a
>> wasted effort.
>>
>> So, I envision 3 possible scenarios:
>>
>> (1) just instrument Mach IR and be done with it;
>>
>> (2) (a) start with Mach IR;
>> (b) later it turns out that extensive portions of
>> (macro)assembly code have to me instrumented (or, for example,
>> C1/Interpreter)
>> (c) implement MacroAssembler mitigations
>>
>> (3) start with MacroAssembler mitigations and be done with it
>> * doesn't perclude gradual roll out across different subsystems
>>
>> Mach IR instrumentation (#1/#2) is the safest variant, but it may
>> require more work.
>>
>> #3 is broadly applicable, but also riskier.
>>
>> What I don't consider as a viable option is C2 CFG instrumentation
>> accompanied by numerous per-instruction mitigations scattered across
>> the code base.
>>
>>>>> I have made a prototype, what this might look like and it looks
>>>>> like this:
>>>>> http://cr.openjdk.java.net/~eosterlund/8234160/webrev.01/
>>>>
>>>> Just one more comment: it's weird to see intel_jcc_erratum
>>>> referenced in shared code. You could #ifdef it for x86-only, but
>>>> it's much better to move the code to x86-specific location.
>>>
>>> Sure, I can move that to an x86 file and make it build only on x86_64.
>>
>> Yes, sounds good. But let's agree on general direction first.
>>
>> Best regards,
>> Vladimir Ivanov
>>
>> [1]
>> http://hg.openjdk.java.net/jdk/jdk/file/tip/src/hotspot/cpu/x86/macroAssembler_x86.hpp#l1666
>>
>>
>> [2]
>> http://hg.openjdk.java.net/jdk/jdk/file/623722a6aeb9/src/hotspot/cpu/x86/stubGenerator_x86_64.cpp
>>
>>
>> [3] http://hg.openjdk.java.net/jdk/jdk/file/tip/src/hotspot/cpu/x86/
>> macroAssembler_x86_(sin|cos|...).cpp
>>
>
More information about the hotspot-compiler-dev
mailing list