/hg/release/icedtea7-forest-2.3/hotspot: 8 new changesets
andrew at icedtea.classpath.org
andrew at icedtea.classpath.org
Fri Mar 21 19:39:28 UTC 2014
changeset 3f0dff7a9cf5 in /hg/release/icedtea7-forest-2.3/hotspot
details: http://icedtea.classpath.org/hg/release/icedtea7-forest-2.3/hotspot?cmd=changeset;node=3f0dff7a9cf5
author: iveresov
date: Mon Mar 12 13:12:07 2012 -0700
7151089: PS NUMA: NUMA allocator should not attempt to free pages when using SHM large pages
Summary: Don't attempt to uncommit SHM-based large pages
Reviewed-by: kvn
changeset cbdd11a54b82 in /hg/release/icedtea7-forest-2.3/hotspot
details: http://icedtea.classpath.org/hg/release/icedtea7-forest-2.3/hotspot?cmd=changeset;node=cbdd11a54b82
author: dcubed
date: Mon Mar 10 18:09:45 2014 +0000
8013057: assert(_needs_gc || SafepointSynchronize::is_at_safepoint()) failed: only read at safepoint
Summary: Detect mmap() commit failures in Linux and Solaris os::commit_memory() impls and call vm_exit_out_of_memory(). Add os::commit_memory_or_exit(). Also tidy up some NMT accounting and some mmap() return value checking.
Reviewed-by: zgu, stefank, dholmes, dsamersoff
changeset 24391b7e49b5 in /hg/release/icedtea7-forest-2.3/hotspot
details: http://icedtea.classpath.org/hg/release/icedtea7-forest-2.3/hotspot?cmd=changeset;node=24391b7e49b5
author: aeriksso
date: Mon Mar 10 18:48:19 2014 +0000
8026887: Make issues due to failed large pages allocations easier to debug
Reviewed-by: stefank, mcastegr, poonam
changeset 3b188862d691 in /hg/release/icedtea7-forest-2.3/hotspot
details: http://icedtea.classpath.org/hg/release/icedtea7-forest-2.3/hotspot?cmd=changeset;node=3b188862d691
author: andrew
date: Wed Feb 19 21:08:04 2014 +0000
PR1679: Allow OpenJDK to build on PaX-enabled kernels
changeset 99ed2a7d2f87 in /hg/release/icedtea7-forest-2.3/hotspot
details: http://icedtea.classpath.org/hg/release/icedtea7-forest-2.3/hotspot?cmd=changeset;node=99ed2a7d2f87
author: dsamersoff
date: Mon Mar 10 22:26:05 2014 +0000
8009062: poor performance of JNI AttachCurrentThread after fix for 7017193
Summary: don't re-evaluate stack bounds for main thread before install guard page
Reviewed-by: coleenp, dholmes, dlong
changeset adf1821ed249 in /hg/release/icedtea7-forest-2.3/hotspot
details: http://icedtea.classpath.org/hg/release/icedtea7-forest-2.3/hotspot?cmd=changeset;node=adf1821ed249
author: chrisphi
date: Fri Mar 14 16:53:01 2014 +0000
Allow ARM32 JIT to be disabled
changeset fba9303068df in /hg/release/icedtea7-forest-2.3/hotspot
details: http://icedtea.classpath.org/hg/release/icedtea7-forest-2.3/hotspot?cmd=changeset;node=fba9303068df
author: andrew
date: Thu Jan 23 22:41:21 2014 +0000
PR1653: Support ppc64le via Zero
changeset 383082077d46 in /hg/release/icedtea7-forest-2.3/hotspot
details: http://icedtea.classpath.org/hg/release/icedtea7-forest-2.3/hotspot?cmd=changeset;node=383082077d46
author: andrew
date: Fri Jan 31 21:14:06 2014 +0000
RH1015432: java-1.7.0-openjdk: Fails on PPC with StackOverflowError (revised fix)
Contributed-by: chphilli at redhat.com
diffstat:
make/linux/makefiles/buildtree.make | 4 +
make/linux/makefiles/zeroshark.make | 4 +-
src/os/bsd/vm/os_bsd.cpp | 43 +-
src/os/bsd/vm/perfMemory_bsd.cpp | 4 +-
src/os/linux/vm/os_linux.cpp | 265 +++++++--
src/os/linux/vm/os_linux.hpp | 8 +-
src/os/linux/vm/perfMemory_linux.cpp | 4 +-
src/os/solaris/vm/os_solaris.cpp | 101 +++-
src/os/solaris/vm/os_solaris.hpp | 5 +-
src/os/solaris/vm/perfMemory_solaris.cpp | 4 +-
src/os/windows/vm/os_windows.cpp | 57 +-
src/os/windows/vm/perfMemory_windows.cpp | 4 +-
src/os_cpu/linux_zero/vm/globals_linux_zero.hpp | 6 +-
src/os_cpu/linux_zero/vm/os_linux_zero.cpp | 10 +-
src/share/vm/gc_implementation/parallelScavenge/cardTableExtension.cpp | 10 +-
src/share/vm/gc_implementation/parallelScavenge/psVirtualspace.cpp | 12 +-
src/share/vm/memory/cardTableModRefBS.cpp | 17 +-
src/share/vm/runtime/os.cpp | 39 +-
src/share/vm/runtime/os.hpp | 22 +-
src/share/vm/runtime/virtualspace.cpp | 17 +-
src/share/vm/utilities/vmError.cpp | 12 +
21 files changed, 471 insertions(+), 177 deletions(-)
diffs (truncated from 1188 to 500 lines):
diff -r 3442eb7ef2d2 -r 383082077d46 make/linux/makefiles/buildtree.make
--- a/make/linux/makefiles/buildtree.make Tue Jan 14 20:24:44 2014 -0500
+++ b/make/linux/makefiles/buildtree.make Fri Jan 31 21:14:06 2014 +0000
@@ -435,6 +435,10 @@
echo " GAMMA_PROG=gamma_g"; \
echo "fi"; \
echo ""; \
+ echo "if [ -x \"$(PAX_COMMAND)\" ]; then "; \
+ echo " $(PAX_COMMAND) $(PAX_COMMAND_ARGS) ./\$${GAMMA_PROG}"; \
+ echo "fi"; \
+ echo ""; \
echo "if [ \"$(OS_VENDOR)\" = \"Darwin\" ]; then "; \
echo " # Ensure architecture for gamma and JAVA_HOME is the same."; \
echo " # NOTE: gamma assumes the OpenJDK directory layout."; \
diff -r 3442eb7ef2d2 -r 383082077d46 make/linux/makefiles/zeroshark.make
--- a/make/linux/makefiles/zeroshark.make Tue Jan 14 20:24:44 2014 -0500
+++ b/make/linux/makefiles/zeroshark.make Fri Jan 31 21:14:06 2014 +0000
@@ -26,7 +26,8 @@
# Setup common to Zero (non-Shark) and Shark versions of VM
ifeq ($(ZERO_LIBARCH),arm)
-
+# check to see if we are building the assembler jit or just zero.
+ifeq ($(ARM32JIT),true)
Obj_Files += asm_helper.o
Obj_Files += cppInterpreter_arm.o
Obj_Files += thumb2.o
@@ -54,6 +55,7 @@
$(CC_COMPILE) $(CFLAGS) -DSTATIC_OFFSETS -o $@ $< $(COMPILE_DONE)
endif
+endif
%.o: %.S
@echo Assembling $<
diff -r 3442eb7ef2d2 -r 383082077d46 src/os/bsd/vm/os_bsd.cpp
--- a/src/os/bsd/vm/os_bsd.cpp Tue Jan 14 20:24:44 2014 -0500
+++ b/src/os/bsd/vm/os_bsd.cpp Fri Jan 31 21:14:06 2014 +0000
@@ -2762,6 +2762,13 @@
}
}
+static void warn_fail_commit_memory(char* addr, size_t size, bool exec,
+ int err) {
+ warning("INFO: os::commit_memory(" PTR_FORMAT ", " SIZE_FORMAT
+ ", %d) failed; error='%s' (errno=%d)", addr, size, exec,
+ strerror(err), err);
+}
+
// NOTE: Bsd kernel does not really reserve the pages for us.
// All it does is to check if there are enough free pages
// left at the time of mmap(). This could be a potential
@@ -2770,12 +2777,22 @@
int prot = exec ? PROT_READ|PROT_WRITE|PROT_EXEC : PROT_READ|PROT_WRITE;
#ifdef __OpenBSD__
// XXX: Work-around mmap/MAP_FIXED bug temporarily on OpenBSD
- return ::mprotect(addr, size, prot) == 0;
+ if (::mprotect(addr, size, prot) == 0) {
+ return true;
+ }
#else
uintptr_t res = (uintptr_t) ::mmap(addr, size, prot,
MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0);
- return res != (uintptr_t) MAP_FAILED;
+ if (res != (uintptr_t) MAP_FAILED) {
+ return true;
+ }
#endif
+
+ // Warn about any commit errors we see in non-product builds just
+ // in case mmap() doesn't work as described on the man page.
+ NOT_PRODUCT(warn_fail_commit_memory(addr, size, exec, errno);)
+
+ return false;
}
#ifndef _ALLBSD_SOURCE
@@ -2803,9 +2820,27 @@
}
#endif
+ // alignment_hint is ignored on this OS
return commit_memory(addr, size, exec);
}
+void os::commit_memory_or_exit(char* addr, size_t size, bool exec,
+ const char* mesg) {
+ assert(mesg != NULL, "mesg must be specified");
+ if (!pd_commit_memory(addr, size, exec)) {
+ // add extra info in product mode for vm_exit_out_of_memory():
+ PRODUCT_ONLY(warn_fail_commit_memory(addr, size, exec, errno);)
+ vm_exit_out_of_memory(size, mesg);
+ }
+}
+
+void os::commit_memory_or_exit(char* addr, size_t size,
+ size_t alignment_hint, bool exec,
+ const char* mesg) {
+ // alignment_hint is ignored on this OS
+ pd_commit_memory_or_exit(addr, size, exec, mesg);
+}
+
void os::realign_memory(char *addr, size_t bytes, size_t alignment_hint) {
#ifndef _ALLBSD_SOURCE
if (UseHugeTLBFS && alignment_hint > (size_t)vm_page_size()) {
@@ -2970,7 +3005,7 @@
}
bool os::create_stack_guard_pages(char* addr, size_t size) {
- return os::commit_memory(addr, size);
+ return os::commit_memory(addr, size, !ExecMem);
}
// If this is a growable mapping, remove the guard pages entirely by
@@ -4467,7 +4502,7 @@
if (!UseMembar) {
address mem_serialize_page = (address) ::mmap(NULL, Bsd::page_size(), PROT_READ | PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0);
- guarantee( mem_serialize_page != NULL, "mmap Failed for memory serialize page");
+ guarantee( mem_serialize_page != MAP_FAILED, "mmap Failed for memory serialize page");
os::set_memory_serialize_page( mem_serialize_page );
#ifndef PRODUCT
diff -r 3442eb7ef2d2 -r 383082077d46 src/os/bsd/vm/perfMemory_bsd.cpp
--- a/src/os/bsd/vm/perfMemory_bsd.cpp Tue Jan 14 20:24:44 2014 -0500
+++ b/src/os/bsd/vm/perfMemory_bsd.cpp Fri Jan 31 21:14:06 2014 +0000
@@ -1,5 +1,5 @@
/*
- * Copyright (c) 2001, 2010, Oracle and/or its affiliates. All rights reserved.
+ * Copyright (c) 2001, 2013, Oracle and/or its affiliates. All rights reserved.
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
*
* This code is free software; you can redistribute it and/or modify it
@@ -59,7 +59,7 @@
}
// commit memory
- if (!os::commit_memory(mapAddress, size)) {
+ if (!os::commit_memory(mapAddress, size, !ExecMem)) {
if (PrintMiscellaneous && Verbose) {
warning("Could not commit PerfData memory\n");
}
diff -r 3442eb7ef2d2 -r 383082077d46 src/os/linux/vm/os_linux.cpp
--- a/src/os/linux/vm/os_linux.cpp Tue Jan 14 20:24:44 2014 -0500
+++ b/src/os/linux/vm/os_linux.cpp Fri Jan 31 21:14:06 2014 +0000
@@ -2508,11 +2508,57 @@
}
}
+static bool recoverable_mmap_error(int err) {
+ // See if the error is one we can let the caller handle. This
+ // list of errno values comes from JBS-6843484. I can't find a
+ // Linux man page that documents this specific set of errno
+ // values so while this list currently matches Solaris, it may
+ // change as we gain experience with this failure mode.
+ switch (err) {
+ case EBADF:
+ case EINVAL:
+ case ENOTSUP:
+ // let the caller deal with these errors
+ return true;
+
+ default:
+ // Any remaining errors on this OS can cause our reserved mapping
+ // to be lost. That can cause confusion where different data
+ // structures think they have the same memory mapped. The worst
+ // scenario is if both the VM and a library think they have the
+ // same memory mapped.
+ return false;
+ }
+}
+
+static void warn_fail_commit_memory(char* addr, size_t size, bool exec,
+ int err) {
+ warning("INFO: os::commit_memory(" PTR_FORMAT ", " SIZE_FORMAT
+ ", %d) failed; error='%s' (errno=%d)", addr, size, exec,
+ strerror(err), err);
+}
+
+static void warn_fail_commit_memory(char* addr, size_t size,
+ size_t alignment_hint, bool exec,
+ int err) {
+ warning("INFO: os::commit_memory(" PTR_FORMAT ", " SIZE_FORMAT
+ ", " SIZE_FORMAT ", %d) failed; error='%s' (errno=%d)", addr, size,
+ alignment_hint, exec, strerror(err), err);
+}
+
+static void warn_fail_commit_memory(char* addr, size_t size,
+ size_t alignment_hint, bool exec,
+ int err, const char* msg) {
+ warning("INFO: os::commit_memory(" PTR_FORMAT ", " SIZE_FORMAT
+ ", " SIZE_FORMAT ", %d) failed; error='%s' (errno=%d); %s", addr, size,
+ alignment_hint, exec, strerror(err), err, msg);
+}
+
// NOTE: Linux kernel does not really reserve the pages for us.
// All it does is to check if there are enough free pages
// left at the time of mmap(). This could be a potential
// problem.
-bool os::commit_memory(char* addr, size_t size, bool exec) {
+int os::Linux::commit_memory_impl(char* addr, size_t size, bool exec) {
int prot = exec ? PROT_READ|PROT_WRITE|PROT_EXEC : PROT_READ|PROT_WRITE;
uintptr_t res = (uintptr_t) ::mmap(addr, size, prot,
MAP_PRIVATE|MAP_FIXED|MAP_ANONYMOUS, -1, 0);
@@ -2520,9 +2566,32 @@
if (UseNUMAInterleaving) {
numa_make_global(addr, size);
}
- return true;
- }
- return false;
+ return 0;
+ }
+
+ int err = errno; // save errno from mmap() call above
+
+ if (!recoverable_mmap_error(err)) {
+ warn_fail_commit_memory(addr, size, exec, err);
+ vm_exit_out_of_memory(size, "committing reserved memory.");
+ }
+
+ return err;
+}
+
+bool os::commit_memory(char* addr, size_t size, bool exec) {
+ return os::Linux::commit_memory_impl(addr, size, exec) == 0;
+}
+
+void os::commit_memory_or_exit(char* addr, size_t size, bool exec,
+ const char* mesg) {
+ assert(mesg != NULL, "mesg must be specified");
+ int err = os::Linux::commit_memory_impl(addr, size, exec);
+ if (err != 0) {
+ // the caller wants all commit errors to exit with the specified mesg:
+ warn_fail_commit_memory(addr, size, exec, err);
+ vm_exit_out_of_memory(size, mesg);
+ }
}
// Define MAP_HUGETLB here so we can build HotSpot on old systems.
@@ -2535,8 +2604,11 @@
#define MADV_HUGEPAGE 14
#endif
-bool os::commit_memory(char* addr, size_t size, size_t alignment_hint,
- bool exec) {
+volatile jint os::Linux::num_largepage_commit_fails = 0;
+
+int os::Linux::commit_memory_impl(char* addr, size_t size,
+ size_t alignment_hint, bool exec) {
+ int err;
if (UseHugeTLBFS && alignment_hint > (size_t)vm_page_size()) {
int prot = exec ? PROT_READ|PROT_WRITE|PROT_EXEC : PROT_READ|PROT_WRITE;
uintptr_t res =
@@ -2547,16 +2619,47 @@
if (UseNUMAInterleaving) {
numa_make_global(addr, size);
}
- return true;
+ return 0;
+ }
+
+ err = errno; // save errno from mmap() call above
+
+ if (!recoverable_mmap_error(err)) {
+ // However, it is not clear that this loss of our reserved mapping
+ // happens with large pages on Linux or that we cannot recover
+ // from the loss. For now, we just issue a warning and we don't
+ // call vm_exit_out_of_memory(). This issue is being tracked by
+ // JBS-8007074.
+ Atomic::inc(&os::Linux::num_largepage_commit_fails);
+ warn_fail_commit_memory(addr, size, alignment_hint, exec, err,
+ "Cannot allocate large pages, falling back to regular pages");
+// vm_exit_out_of_memory(size, "committing reserved memory.");
}
// Fall through and try to use small pages
}
- if (commit_memory(addr, size, exec)) {
+ err = os::Linux::commit_memory_impl(addr, size, exec);
+ if (err == 0) {
realign_memory(addr, size, alignment_hint);
- return true;
- }
- return false;
+ }
+ return err;
+}
+
+bool os::commit_memory(char* addr, size_t size, size_t alignment_hint,
+ bool exec) {
+ return os::Linux::commit_memory_impl(addr, size, alignment_hint, exec) == 0;
+}
+
+void os::commit_memory_or_exit(char* addr, size_t size,
+ size_t alignment_hint, bool exec,
+ const char* mesg) {
+ assert(mesg != NULL, "mesg must be specified");
+ int err = os::Linux::commit_memory_impl(addr, size, alignment_hint, exec);
+ if (err != 0) {
+ // the caller wants all commit errors to exit with the specified mesg:
+ warn_fail_commit_memory(addr, size, alignment_hint, exec, err);
+ vm_exit_out_of_memory(size, mesg);
+ }
}
void os::realign_memory(char *addr, size_t bytes, size_t alignment_hint) {
@@ -2568,7 +2671,14 @@
}
void os::free_memory(char *addr, size_t bytes, size_t alignment_hint) {
- commit_memory(addr, bytes, alignment_hint, false);
+ // This method works by doing an mmap over an existing mmaping and effectively discarding
+ // the existing pages. However it won't work for SHM-based large pages that cannot be
+ // uncommitted at all. We don't do anything in this case to avoid creating a segment with
+ // small pages on top of the SHM segment. This method always works for small pages, so we
+ // allow that in any case.
+ if (alignment_hint <= (size_t)os::vm_page_size() || !UseSHM) {
+ commit_memory(addr, bytes, alignment_hint, !ExecMem);
+ }
}
void os::numa_make_global(char *addr, size_t bytes) {
@@ -2751,6 +2861,53 @@
return res != (uintptr_t) MAP_FAILED;
}
+static
+address get_stack_commited_bottom(address bottom, size_t size) {
+ address nbot = bottom;
+ address ntop = bottom + size;
+
+ size_t page_sz = os::vm_page_size();
+ unsigned pages = size / page_sz;
+
+ unsigned char vec[1];
+ unsigned imin = 1, imax = pages + 1, imid;
+ int mincore_return_value;
+
+ while (imin < imax) {
+ imid = (imax + imin) / 2;
+ nbot = ntop - (imid * page_sz);
+
+ // Use a trick with mincore to check whether the page is mapped or not.
+ // mincore sets vec to 1 if page resides in memory and to 0 if page
+ // is swapped output but if page we are asking for is unmapped
+ // it returns -1,ENOMEM
+ mincore_return_value = mincore(nbot, page_sz, vec);
+
+ if (mincore_return_value == -1) {
+ // Page is not mapped go up
+ // to find first mapped page
+ if (errno != EAGAIN) {
+ assert(errno == ENOMEM, "Unexpected mincore errno");
+ imax = imid;
+ }
+ } else {
+ // Page is mapped go down
+ // to find first not mapped page
+ imin = imid + 1;
+ }
+ }
+
+ nbot = nbot + page_sz;
+
+ // Adjust stack bottom one page up if last checked page is not mapped
+ if (mincore_return_value == -1) {
+ nbot = nbot + page_sz;
+ }
+
+ return nbot;
+}
+
+
// Linux uses a growable mapping for the stack, and if the mapping for
// the stack guard pages is not removed when we detach a thread the
// stack cannot grow beyond the pages where the stack guard was
@@ -2765,74 +2922,52 @@
// So, we need to know the extent of the stack mapping when
// create_stack_guard_pages() is called.
-// Find the bounds of the stack mapping. Return true for success.
-//
// We only need this for stacks that are growable: at the time of
// writing thread stacks don't use growable mappings (i.e. those
// creeated with MAP_GROWSDOWN), and aren't marked "[stack]", so this
// only applies to the main thread.
-static
-bool get_stack_bounds(uintptr_t *bottom, uintptr_t *top) {
-
- char buf[128];
- int fd, sz;
-
- if ((fd = ::open("/proc/self/maps", O_RDONLY)) < 0) {
- return false;
- }
-
- const char kw[] = "[stack]";
- const int kwlen = sizeof(kw)-1;
-
- // Address part of /proc/self/maps couldn't be more than 128 bytes
- while ((sz = os::get_line_chars(fd, buf, sizeof(buf))) > 0) {
- if (sz > kwlen && ::memcmp(buf+sz-kwlen, kw, kwlen) == 0) {
- // Extract addresses
- if (sscanf(buf, "%" SCNxPTR "-%" SCNxPTR, bottom, top) == 2) {
- uintptr_t sp = (uintptr_t) __builtin_frame_address(0);
- if (sp >= *bottom && sp <= *top) {
- ::close(fd);
- return true;
- }
- }
- }
- }
-
- ::close(fd);
- return false;
-}
-
-
// If the (growable) stack mapping already extends beyond the point
// where we're going to put our guard pages, truncate the mapping at
// that point by munmap()ping it. This ensures that when we later
// munmap() the guard pages we don't leave a hole in the stack
-// mapping. This only affects the main/initial thread, but guard
-// against future OS changes
+// mapping. This only affects the main/initial thread
+
bool os::create_stack_guard_pages(char* addr, size_t size) {
- uintptr_t stack_extent, stack_base;
- bool chk_bounds = NOT_DEBUG(os::Linux::is_initial_thread()) DEBUG_ONLY(true);
- if (chk_bounds && get_stack_bounds(&stack_extent, &stack_base)) {
- assert(os::Linux::is_initial_thread(),
- "growable stack in non-initial thread");
- if (stack_extent < (uintptr_t)addr)
- ::munmap((void*)stack_extent, (uintptr_t)addr - stack_extent);
- }
-
- return os::commit_memory(addr, size);
+
+ if (os::Linux::is_initial_thread()) {
+ // As we manually grow stack up to bottom inside create_attached_thread(),
+ // it's likely that os::Linux::initial_thread_stack_bottom is mapped and
+ // we don't need to do anything special.
+ // Check it first, before calling heavy function.
+ uintptr_t stack_extent = (uintptr_t) os::Linux::initial_thread_stack_bottom();
+ unsigned char vec[1];
+
+ if (mincore((address)stack_extent, os::vm_page_size(), vec) == -1) {
+ // Fallback to slow path on all errors, including EAGAIN
+ stack_extent = (uintptr_t) get_stack_commited_bottom(
+ os::Linux::initial_thread_stack_bottom(),
+ (size_t)addr - stack_extent);
+ }
+
+ if (stack_extent < (uintptr_t)addr) {
+ ::munmap((void*)stack_extent, (uintptr_t)(addr - stack_extent));
+ }
+ }
+
+ return os::commit_memory(addr, size, !ExecMem);
}
// If this is a growable mapping, remove the guard pages entirely by
// munmap()ping them. If not, just call uncommit_memory(). This only
// affects the main/initial thread, but guard against future OS changes
+// It's safe to always unmap guard pages for initial thread because we
+// always place it right after end of the mapped region
+
bool os::remove_stack_guard_pages(char* addr, size_t size) {
uintptr_t stack_extent, stack_base;
- bool chk_bounds = NOT_DEBUG(os::Linux::is_initial_thread()) DEBUG_ONLY(true);
- if (chk_bounds && get_stack_bounds(&stack_extent, &stack_base)) {
- assert(os::Linux::is_initial_thread(),
- "growable stack in non-initial thread");
-
+
+ if (os::Linux::is_initial_thread()) {
return ::munmap(addr, size) == 0;
}
@@ -2941,7 +3076,7 @@
MAP_ANONYMOUS|MAP_PRIVATE|MAP_HUGETLB,
-1, 0);
- if (p != (void *) -1) {
+ if (p != MAP_FAILED) {
// We don't know if this really is a huge page or not.
FILE *fp = fopen("/proc/self/maps", "r");
if (fp) {
@@ -4218,7 +4353,7 @@
if (!UseMembar) {
address mem_serialize_page = (address) ::mmap(NULL, Linux::page_size(), PROT_READ | PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0);
- guarantee( mem_serialize_page != NULL, "mmap Failed for memory serialize page");
+ guarantee( mem_serialize_page != MAP_FAILED, "mmap Failed for memory serialize page");
os::set_memory_serialize_page( mem_serialize_page );
#ifndef PRODUCT
diff -r 3442eb7ef2d2 -r 383082077d46 src/os/linux/vm/os_linux.hpp
More information about the distro-pkg-dev
mailing list