premain: performance sniff tests
Vladimir Ivanov
vladimir.x.ivanov at oracle.com
Tue Sep 5 23:23:41 UTC 2023
Hi Ashutosh,
Thanks for giving it a try!
There were some experiments with PetClinic on our side before and it was
noticed that the application relies on custom loaders which aren't fully
supported yet. It was the main limiting factor for new optimizations.
Until proper support for custom loaders is there, I suggest to modify
the benchmark so it relies only on existing system loaders.
Speaking of peak performance, some loss of performance is expected.
Cached code is compiled conservatively (e.g., no constant folding for
static final fields) so it can be reused in deployment runs. For now,
the intended solution is to eventually recompile cached code online with
all the optimizations enabled (have to be explicitly enabled
-XX:+UseRecompilation). It's a work-in-progress and our experience using
it was mixed: recompilation doesn't always fully restore peak
performance.
But assuming that both CDS and cached code archive are underutilized
(due to aforementioned reliance on custom loaders), 10% sounds way too
big of a difference. I suggest to experiment with different flag
combinations (e.g., turning ReplayTraining and LoadCachedCode on and off
independently).
There's additional diagnostic output JVM produces which may help to
observe effects from new optimizations during both training and
deployment runs:
* -XX:+PrintCompilation: compilations satisfied from cached code
archive are marked w/ "R";
* -XX:+CITime: prints information about cached code archive usage;
* -Xlog:init=info: produces additional information about some startup
activities
* -XX:+PrintSharedArchiveAndExit additionally dumps training data and
cached code archive info
* -Xlog:scc*=info and -Xlog:cds*=info print lots of additional
information both during training and deployment
Hope it helps.
Best regards,
Vladimir Ivanov
On 9/5/23 13:52, Ashutosh Mehra wrote:
> Hi,
>
> We have been interested in persisting the profiling data in the CDS
> archive with the intention of improving the application's warmup time.
> And now that the premain branch is here that does save profile data
> along with AOT, we started playing with the premain branch to understand
> its impact on the performance.
>
> Our setup uses Springboot Petclinic [0] application and the CDS and
> shared code archives are generated in a manner similar to this script [1].
> Our training run only covers the application startup phase. That means
> at each step we start the application and shut it down without putting
> any load on it.
>
> Using the archives thus generated I have done few experiments on my
> local system. In these experiments the application is bound to two cpus.
> The baseline for comparing the results is the case where the CDS archive
> does not have any profiling data and there is no shared code archive.
> The "premain" configuration refers to using a shared code archive and a
> CDS archive with training data.
>
> Here are some initial results:
>
> 1. Startup: It is heartening to see start-up time improve by almost 11%.
>
> baseline 10.2s
> premain 9.1s
>
> 2. Warmup:
> This test measures the warmup time by applying load using 1 jmeter
> thread to get an idea of the ramp-up time to reach the peak throughput.
> The load is applied for the duration of 300 seconds. The graph [2] for
> aot+profiling configuration shows interesting behavior.
> In the initial period premain is ramping up faster than the baseline.
> Then the slope of the curve for premain reduces significantly and a
> couple of dips are also seen. Finally the throughput stabilizes.
> It shows a drastic difference in the warmup time of the application when
> running with the "premain" config.
>
> 3. Peak throughput: Last experiment is to measure peak throughput. It
> starts with a warm-up phase of 180 seconds using 1 jmeter thread. After
> the warmup phase the load is applied with 10 jmeter threads for a
> duration of 5 mins.
> Last two minutes of throughput is considered for measurement. The graph
> [3] for this test shows almost a 10% drop in the throughput compared to
> the baseline.
>
>
> I am sure others would have done similar testing. My questions are:
>
> 1. Are these results on the expected lines?
> 2. Are these tests using the CDS and the shared code (or cached code)
> archives in the expected manner.
> 3. Warmup time with the premain branch looks pretty bad which is
> surprising. Is there any trick I missed in my tests? Is there anything
> else that needs to be done to get better warmup time?
> 4. What is the point of creating a new static archive? Shouldn't the
> applications just create the dynamic archive?
> 5. I am also wondering if there is any design doc that can be shared
> that explains the AOT compilation strategy adopted in the premain branch?
>
> I have placed my scripts here [4] in case anyone wants to use them to
> run these tests (you need to build the Petclinic app before using these
> scripts).
>
> Please feel free to share your thoughts.
>
> [0] https://github.com/spring-projects/spring-petclinic
> <https://github.com/spring-projects/spring-petclinic>
> [1]
> https://github.com/openjdk/leyden/blob/d960fb15258cc99a1bf7f0b1e94bd8be06605aad/test/hotspot/jtreg/premain/lib/premain-run.sh#L70-L101 <https://github.com/openjdk/leyden/blob/d960fb15258cc99a1bf7f0b1e94bd8be06605aad/test/hotspot/jtreg/premain/lib/premain-run.sh#L70-L101>
> [2]
> https://github.com/ashu-mehra/leyden-perf/blob/main/spring/fd82682/tput-t1.svg <https://github.com/ashu-mehra/leyden-perf/blob/main/spring/fd82682/tput-t1.svg>
> [3]
> https://github.com/ashu-mehra/leyden-perf/blob/main/spring/fd82682/tput-t10.svg <https://github.com/ashu-mehra/leyden-perf/blob/main/spring/fd82682/tput-t10.svg>
> [4] https://github.com/ashu-mehra/leyden-perf
> <https://github.com/ashu-mehra/leyden-perf>
>
> Thanks,
> - Ashutosh Mehra
More information about the leyden-dev
mailing list