Review Request: UseNUMAInterleaving

Deneau, Tom tom.deneau at amd.com
Mon May 16 10:54:11 PDT 2011


Please review this patch which adds a new flag called
UseNUMAInterleaving.  This flag provides a subset of the functionality
provided by UseNUMA, and its main purpose is to provide that subset on
OSes like Windows which do not support the full UseNUMA functionality.
In UseNUMA terminology, UseNUMAInterleaved makes all memory
"numa_global" which is implemented as interleaved.

The situations where this shows the biggest benefits would be:
   * Windows platforms with multiple numa nodes (eg, 4)
 
   * The JVM process is run across all the nodes (not affinitized to one node).

   * A workload that uses the majority of the cores in the machine, so
     that the heap is being accessed from many cores, including remote
     ones.

   * Enough memory per node and a heap size such that the default heap
     placement policy on windows would end up with the heap (or
     nursery) placed on one node.

jbb2005 and SPECPower_ssj2008 are examples of such workloads.  In our
measurements, we have seen some cases where the performance with
UseNUMAInterleaving was 2.7x vs. the performance without. There were
gains of varying sizes across all systems.

As currently implemented this flag is ignored on Linux and Solaris
since they already support the full UseNUMA flag.

The webrev is at
http://cr.openjdk.java.net/~tdeneau/UseNUMAInterleaving/webrev.01/

Summary of changes:
 
   * Other than adding the new UseNUMAInterleaving global flag, all of
     the changes are in src/os/windows/vm/os_windows.cpp

   * Some static routines were added to set things up init time.  These
      * check that the required APIs (VirtualAllocExNuma,
        GetNumaHighestNodeNumber, GetNumaNodeProcessorMask) exist in
        the OS

      * build the list of numa nodes on which this process has affinity

   * Changes to os::reserve_memory
      * There was already a routine that reserved pages one page at a
        time (used for Individual Large Page Allocation on WS2003).
        This was abstracted to a separate routine, called
        allocate_pages_individually.  This gets called both for the
        Individual Large Page Allocation thing mentioned above and for
        UseNUMAInterleaving (for both small and large pages)

      * When used for NUMA Interleaving this just goes thru the numa
        node list in a round-robin fashion, using a different one for
        each chunk (with 4K pages, the minimum allocation granularity
        is 64K, with 2M pages it is 1 Page)

      * Whether we do just a reserve or a combined reserve/commit is
        determined by the caller of allocate_pages_individually

         * When used with large pages, we do a Reserve and Commit at
           the same time which is the way it always worked and the way
           it has to work on windows.

         * For small pages, only the reserve is done, the commit will
           come later. (which is the way it worked for
           non-interleaved)

   * os::commit_memory changes
      * If UseNUMAIntereaving is true, os::commit_memory has to check
        whether it was being asked to commit memory that might have
        come from multiple Reserve allocations, if so, the commits
        must also be broken up.  We don't keep any data structure to
        keep track of this, we just use VirtualQuery which queries the
        properties of a VA range and can tell us how much came from
        one VirtualAlloc call.

I do not have a bug id for this.

-- Tom Deneau, AMD



More information about the hotspot-compiler-dev mailing list