Wed, 15 May 2013 11:05:09 +0200
6843347: Boundary values in some public GC options cause crashes
Summary: Setting some public integer options to specific values causes crashes or undefined GC behavior. This patchset adds the necessary argument checking for these options.
Reviewed-by: jmasa, brutisso
1 /*
2 * Copyright (c) 1997, 2013, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
25 #ifndef SHARE_VM_MEMORY_ALLOCATION_INLINE_HPP
26 #define SHARE_VM_MEMORY_ALLOCATION_INLINE_HPP
28 #include "runtime/atomic.inline.hpp"
29 #include "runtime/os.hpp"
31 // Explicit C-heap memory management
33 void trace_heap_malloc(size_t size, const char* name, void *p);
34 void trace_heap_free(void *p);
36 #ifndef PRODUCT
37 // Increments unsigned long value for statistics (not atomic on MP).
38 inline void inc_stat_counter(volatile julong* dest, julong add_value) {
39 #if defined(SPARC) || defined(X86)
40 // Sparc and X86 have atomic jlong (8 bytes) instructions
41 julong value = Atomic::load((volatile jlong*)dest);
42 value += add_value;
43 Atomic::store((jlong)value, (volatile jlong*)dest);
44 #else
45 // possible word-tearing during load/store
46 *dest += add_value;
47 #endif
48 }
49 #endif
51 // allocate using malloc; will fail if no memory available
52 inline char* AllocateHeap(size_t size, MEMFLAGS flags, address pc = 0,
53 AllocFailType alloc_failmode = AllocFailStrategy::EXIT_OOM) {
54 if (pc == 0) {
55 pc = CURRENT_PC;
56 }
57 char* p = (char*) os::malloc(size, flags, pc);
58 #ifdef ASSERT
59 if (PrintMallocFree) trace_heap_malloc(size, "AllocateHeap", p);
60 #endif
61 if (p == NULL && alloc_failmode == AllocFailStrategy::EXIT_OOM) {
62 vm_exit_out_of_memory(size, OOM_MALLOC_ERROR, "AllocateHeap");
63 }
64 return p;
65 }
67 inline char* ReallocateHeap(char *old, size_t size, MEMFLAGS flags,
68 AllocFailType alloc_failmode = AllocFailStrategy::EXIT_OOM) {
69 char* p = (char*) os::realloc(old, size, flags, CURRENT_PC);
70 #ifdef ASSERT
71 if (PrintMallocFree) trace_heap_malloc(size, "ReallocateHeap", p);
72 #endif
73 if (p == NULL && alloc_failmode == AllocFailStrategy::EXIT_OOM) {
74 vm_exit_out_of_memory(size, OOM_MALLOC_ERROR, "ReallocateHeap");
75 }
76 return p;
77 }
79 inline void FreeHeap(void* p, MEMFLAGS memflags = mtInternal) {
80 #ifdef ASSERT
81 if (PrintMallocFree) trace_heap_free(p);
82 #endif
83 os::free(p, memflags);
84 }
87 template <MEMFLAGS F> void* CHeapObj<F>::operator new(size_t size,
88 address caller_pc){
89 #ifdef ASSERT
90 void* p = (void*)AllocateHeap(size, F, (caller_pc != 0 ? caller_pc : CALLER_PC));
91 if (PrintMallocFree) trace_heap_malloc(size, "CHeapObj-new", p);
92 return p;
93 #else
94 return (void *) AllocateHeap(size, F, (caller_pc != 0 ? caller_pc : CALLER_PC));
95 #endif
96 }
98 template <MEMFLAGS F> void* CHeapObj<F>::operator new (size_t size,
99 const std::nothrow_t& nothrow_constant, address caller_pc) {
100 #ifdef ASSERT
101 void* p = (void*)AllocateHeap(size, F, (caller_pc != 0 ? caller_pc : CALLER_PC),
102 AllocFailStrategy::RETURN_NULL);
103 if (PrintMallocFree) trace_heap_malloc(size, "CHeapObj-new", p);
104 return p;
105 #else
106 return (void *) AllocateHeap(size, F, (caller_pc != 0 ? caller_pc : CALLER_PC),
107 AllocFailStrategy::RETURN_NULL);
108 #endif
109 }
111 template <MEMFLAGS F> void CHeapObj<F>::operator delete(void* p){
112 FreeHeap(p, F);
113 }
115 template <class E, MEMFLAGS F>
116 E* ArrayAllocator<E, F>::allocate(size_t length) {
117 assert(_addr == NULL, "Already in use");
119 _size = sizeof(E) * length;
120 _use_malloc = _size < ArrayAllocatorMallocLimit;
122 if (_use_malloc) {
123 _addr = AllocateHeap(_size, F);
124 if (_addr == NULL && _size >= (size_t)os::vm_allocation_granularity()) {
125 // malloc failed let's try with mmap instead
126 _use_malloc = false;
127 } else {
128 return (E*)_addr;
129 }
130 }
132 int alignment = os::vm_allocation_granularity();
133 _size = align_size_up(_size, alignment);
135 _addr = os::reserve_memory(_size, NULL, alignment, F);
136 if (_addr == NULL) {
137 vm_exit_out_of_memory(_size, OOM_MMAP_ERROR, "Allocator (reserve)");
138 }
140 bool success = os::commit_memory(_addr, _size, false /* executable */);
141 if (!success) {
142 vm_exit_out_of_memory(_size, OOM_MMAP_ERROR, "Allocator (commit)");
143 }
145 return (E*)_addr;
146 }
148 template<class E, MEMFLAGS F>
149 void ArrayAllocator<E, F>::free() {
150 if (_addr != NULL) {
151 if (_use_malloc) {
152 FreeHeap(_addr, F);
153 } else {
154 os::release_memory(_addr, _size);
155 }
156 _addr = NULL;
157 }
158 }
160 #endif // SHARE_VM_MEMORY_ALLOCATION_INLINE_HPP