Thu, 27 Dec 2018 11:43:33 +0800
Merge
aoqi@0 | 1 | /* |
aoqi@0 | 2 | * Copyright (c) 2009, 2012, Oracle and/or its affiliates. All rights reserved. |
aoqi@0 | 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
aoqi@0 | 4 | * |
aoqi@0 | 5 | * This code is free software; you can redistribute it and/or modify it |
aoqi@0 | 6 | * under the terms of the GNU General Public License version 2 only, as |
aoqi@0 | 7 | * published by the Free Software Foundation. |
aoqi@0 | 8 | * |
aoqi@0 | 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
aoqi@0 | 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
aoqi@0 | 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
aoqi@0 | 12 | * version 2 for more details (a copy is included in the LICENSE file that |
aoqi@0 | 13 | * accompanied this code). |
aoqi@0 | 14 | * |
aoqi@0 | 15 | * You should have received a copy of the GNU General Public License version |
aoqi@0 | 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
aoqi@0 | 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
aoqi@0 | 18 | * |
aoqi@0 | 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
aoqi@0 | 20 | * or visit www.oracle.com if you need additional information or have any |
aoqi@0 | 21 | * questions. |
aoqi@0 | 22 | * |
aoqi@0 | 23 | */ |
aoqi@0 | 24 | |
aoqi@0 | 25 | #ifndef SHARE_VM_UTILITIES_STACK_INLINE_HPP |
aoqi@0 | 26 | #define SHARE_VM_UTILITIES_STACK_INLINE_HPP |
aoqi@0 | 27 | |
aoqi@0 | 28 | #include "utilities/stack.hpp" |
aoqi@0 | 29 | |
aoqi@0 | 30 | template <MEMFLAGS F> StackBase<F>::StackBase(size_t segment_size, size_t max_cache_size, |
aoqi@0 | 31 | size_t max_size): |
aoqi@0 | 32 | _seg_size(segment_size), |
aoqi@0 | 33 | _max_cache_size(max_cache_size), |
aoqi@0 | 34 | _max_size(adjust_max_size(max_size, segment_size)) |
aoqi@0 | 35 | { |
aoqi@0 | 36 | assert(_max_size % _seg_size == 0, "not a multiple"); |
aoqi@0 | 37 | } |
aoqi@0 | 38 | |
aoqi@0 | 39 | template <MEMFLAGS F> size_t StackBase<F>::adjust_max_size(size_t max_size, size_t seg_size) |
aoqi@0 | 40 | { |
aoqi@0 | 41 | assert(seg_size > 0, "cannot be 0"); |
aoqi@0 | 42 | assert(max_size >= seg_size || max_size == 0, "max_size too small"); |
aoqi@0 | 43 | const size_t limit = max_uintx - (seg_size - 1); |
aoqi@0 | 44 | if (max_size == 0 || max_size > limit) { |
aoqi@0 | 45 | max_size = limit; |
aoqi@0 | 46 | } |
aoqi@0 | 47 | return (max_size + seg_size - 1) / seg_size * seg_size; |
aoqi@0 | 48 | } |
aoqi@0 | 49 | |
aoqi@0 | 50 | template <class E, MEMFLAGS F> |
aoqi@0 | 51 | Stack<E, F>::Stack(size_t segment_size, size_t max_cache_size, size_t max_size): |
aoqi@0 | 52 | StackBase<F>(adjust_segment_size(segment_size), max_cache_size, max_size) |
aoqi@0 | 53 | { |
aoqi@0 | 54 | reset(true); |
aoqi@0 | 55 | } |
aoqi@0 | 56 | |
aoqi@0 | 57 | template <class E, MEMFLAGS F> |
aoqi@0 | 58 | void Stack<E, F>::push(E item) |
aoqi@0 | 59 | { |
aoqi@0 | 60 | assert(!is_full(), "pushing onto a full stack"); |
aoqi@0 | 61 | if (this->_cur_seg_size == this->_seg_size) { |
aoqi@0 | 62 | push_segment(); |
aoqi@0 | 63 | } |
aoqi@0 | 64 | this->_cur_seg[this->_cur_seg_size] = item; |
aoqi@0 | 65 | ++this->_cur_seg_size; |
aoqi@0 | 66 | } |
aoqi@0 | 67 | |
aoqi@0 | 68 | template <class E, MEMFLAGS F> |
aoqi@0 | 69 | E Stack<E, F>::pop() |
aoqi@0 | 70 | { |
aoqi@0 | 71 | assert(!is_empty(), "popping from an empty stack"); |
aoqi@0 | 72 | if (this->_cur_seg_size == 1) { |
aoqi@0 | 73 | E tmp = _cur_seg[--this->_cur_seg_size]; |
aoqi@0 | 74 | pop_segment(); |
aoqi@0 | 75 | return tmp; |
aoqi@0 | 76 | } |
aoqi@0 | 77 | return this->_cur_seg[--this->_cur_seg_size]; |
aoqi@0 | 78 | } |
aoqi@0 | 79 | |
aoqi@0 | 80 | template <class E, MEMFLAGS F> |
aoqi@0 | 81 | void Stack<E, F>::clear(bool clear_cache) |
aoqi@0 | 82 | { |
aoqi@0 | 83 | free_segments(_cur_seg); |
aoqi@0 | 84 | if (clear_cache) free_segments(_cache); |
aoqi@0 | 85 | reset(clear_cache); |
aoqi@0 | 86 | } |
aoqi@0 | 87 | |
aoqi@0 | 88 | template <class E, MEMFLAGS F> |
aoqi@0 | 89 | size_t Stack<E, F>::adjust_segment_size(size_t seg_size) |
aoqi@0 | 90 | { |
aoqi@0 | 91 | const size_t elem_sz = sizeof(E); |
aoqi@0 | 92 | const size_t ptr_sz = sizeof(E*); |
aoqi@0 | 93 | assert(elem_sz % ptr_sz == 0 || ptr_sz % elem_sz == 0, "bad element size"); |
aoqi@0 | 94 | if (elem_sz < ptr_sz) { |
aoqi@0 | 95 | return align_size_up(seg_size * elem_sz, ptr_sz) / elem_sz; |
aoqi@0 | 96 | } |
aoqi@0 | 97 | return seg_size; |
aoqi@0 | 98 | } |
aoqi@0 | 99 | |
aoqi@0 | 100 | template <class E, MEMFLAGS F> |
aoqi@0 | 101 | size_t Stack<E, F>::link_offset() const |
aoqi@0 | 102 | { |
aoqi@0 | 103 | return align_size_up(this->_seg_size * sizeof(E), sizeof(E*)); |
aoqi@0 | 104 | } |
aoqi@0 | 105 | |
aoqi@0 | 106 | template <class E, MEMFLAGS F> |
aoqi@0 | 107 | size_t Stack<E, F>::segment_bytes() const |
aoqi@0 | 108 | { |
aoqi@0 | 109 | return link_offset() + sizeof(E*); |
aoqi@0 | 110 | } |
aoqi@0 | 111 | |
aoqi@0 | 112 | template <class E, MEMFLAGS F> |
aoqi@0 | 113 | E** Stack<E, F>::link_addr(E* seg) const |
aoqi@0 | 114 | { |
aoqi@0 | 115 | return (E**) ((char*)seg + link_offset()); |
aoqi@0 | 116 | } |
aoqi@0 | 117 | |
aoqi@0 | 118 | template <class E, MEMFLAGS F> |
aoqi@0 | 119 | E* Stack<E, F>::get_link(E* seg) const |
aoqi@0 | 120 | { |
aoqi@0 | 121 | return *link_addr(seg); |
aoqi@0 | 122 | } |
aoqi@0 | 123 | |
aoqi@0 | 124 | template <class E, MEMFLAGS F> |
aoqi@0 | 125 | E* Stack<E, F>::set_link(E* new_seg, E* old_seg) |
aoqi@0 | 126 | { |
aoqi@0 | 127 | *link_addr(new_seg) = old_seg; |
aoqi@0 | 128 | return new_seg; |
aoqi@0 | 129 | } |
aoqi@0 | 130 | |
aoqi@0 | 131 | template <class E, MEMFLAGS F> |
aoqi@0 | 132 | E* Stack<E, F>::alloc(size_t bytes) |
aoqi@0 | 133 | { |
aoqi@0 | 134 | return (E*) NEW_C_HEAP_ARRAY(char, bytes, F); |
aoqi@0 | 135 | } |
aoqi@0 | 136 | |
aoqi@0 | 137 | template <class E, MEMFLAGS F> |
aoqi@0 | 138 | void Stack<E, F>::free(E* addr, size_t bytes) |
aoqi@0 | 139 | { |
aoqi@0 | 140 | FREE_C_HEAP_ARRAY(char, (char*) addr, F); |
aoqi@0 | 141 | } |
aoqi@0 | 142 | |
aoqi@0 | 143 | template <class E, MEMFLAGS F> |
aoqi@0 | 144 | void Stack<E, F>::push_segment() |
aoqi@0 | 145 | { |
aoqi@0 | 146 | assert(this->_cur_seg_size == this->_seg_size, "current segment is not full"); |
aoqi@0 | 147 | E* next; |
aoqi@0 | 148 | if (this->_cache_size > 0) { |
aoqi@0 | 149 | // Use a cached segment. |
aoqi@0 | 150 | next = _cache; |
aoqi@0 | 151 | _cache = get_link(_cache); |
aoqi@0 | 152 | --this->_cache_size; |
aoqi@0 | 153 | } else { |
aoqi@0 | 154 | next = alloc(segment_bytes()); |
aoqi@0 | 155 | DEBUG_ONLY(zap_segment(next, true);) |
aoqi@0 | 156 | } |
aoqi@0 | 157 | const bool at_empty_transition = is_empty(); |
aoqi@0 | 158 | this->_cur_seg = set_link(next, _cur_seg); |
aoqi@0 | 159 | this->_cur_seg_size = 0; |
aoqi@0 | 160 | this->_full_seg_size += at_empty_transition ? 0 : this->_seg_size; |
aoqi@0 | 161 | DEBUG_ONLY(verify(at_empty_transition);) |
aoqi@0 | 162 | } |
aoqi@0 | 163 | |
aoqi@0 | 164 | template <class E, MEMFLAGS F> |
aoqi@0 | 165 | void Stack<E, F>::pop_segment() |
aoqi@0 | 166 | { |
aoqi@0 | 167 | assert(this->_cur_seg_size == 0, "current segment is not empty"); |
aoqi@0 | 168 | E* const prev = get_link(_cur_seg); |
aoqi@0 | 169 | if (this->_cache_size < this->_max_cache_size) { |
aoqi@0 | 170 | // Add the current segment to the cache. |
aoqi@0 | 171 | DEBUG_ONLY(zap_segment(_cur_seg, false);) |
aoqi@0 | 172 | _cache = set_link(_cur_seg, _cache); |
aoqi@0 | 173 | ++this->_cache_size; |
aoqi@0 | 174 | } else { |
aoqi@0 | 175 | DEBUG_ONLY(zap_segment(_cur_seg, true);) |
aoqi@0 | 176 | free(_cur_seg, segment_bytes()); |
aoqi@0 | 177 | } |
aoqi@0 | 178 | const bool at_empty_transition = prev == NULL; |
aoqi@0 | 179 | this->_cur_seg = prev; |
aoqi@0 | 180 | this->_cur_seg_size = this->_seg_size; |
aoqi@0 | 181 | this->_full_seg_size -= at_empty_transition ? 0 : this->_seg_size; |
aoqi@0 | 182 | DEBUG_ONLY(verify(at_empty_transition);) |
aoqi@0 | 183 | } |
aoqi@0 | 184 | |
aoqi@0 | 185 | template <class E, MEMFLAGS F> |
aoqi@0 | 186 | void Stack<E, F>::free_segments(E* seg) |
aoqi@0 | 187 | { |
aoqi@0 | 188 | const size_t bytes = segment_bytes(); |
aoqi@0 | 189 | while (seg != NULL) { |
aoqi@0 | 190 | E* const prev = get_link(seg); |
aoqi@0 | 191 | free(seg, bytes); |
aoqi@0 | 192 | seg = prev; |
aoqi@0 | 193 | } |
aoqi@0 | 194 | } |
aoqi@0 | 195 | |
aoqi@0 | 196 | template <class E, MEMFLAGS F> |
aoqi@0 | 197 | void Stack<E, F>::reset(bool reset_cache) |
aoqi@0 | 198 | { |
aoqi@0 | 199 | this->_cur_seg_size = this->_seg_size; // So push() will alloc a new segment. |
aoqi@0 | 200 | this->_full_seg_size = 0; |
aoqi@0 | 201 | _cur_seg = NULL; |
aoqi@0 | 202 | if (reset_cache) { |
aoqi@0 | 203 | this->_cache_size = 0; |
aoqi@0 | 204 | _cache = NULL; |
aoqi@0 | 205 | } |
aoqi@0 | 206 | } |
aoqi@0 | 207 | |
aoqi@0 | 208 | #ifdef ASSERT |
aoqi@0 | 209 | template <class E, MEMFLAGS F> |
aoqi@0 | 210 | void Stack<E, F>::verify(bool at_empty_transition) const |
aoqi@0 | 211 | { |
aoqi@0 | 212 | assert(size() <= this->max_size(), "stack exceeded bounds"); |
aoqi@0 | 213 | assert(this->cache_size() <= this->max_cache_size(), "cache exceeded bounds"); |
aoqi@0 | 214 | assert(this->_cur_seg_size <= this->segment_size(), "segment index exceeded bounds"); |
aoqi@0 | 215 | |
aoqi@0 | 216 | assert(this->_full_seg_size % this->_seg_size == 0, "not a multiple"); |
aoqi@0 | 217 | assert(at_empty_transition || is_empty() == (size() == 0), "mismatch"); |
aoqi@0 | 218 | assert((_cache == NULL) == (this->cache_size() == 0), "mismatch"); |
aoqi@0 | 219 | |
aoqi@0 | 220 | if (is_empty()) { |
aoqi@0 | 221 | assert(this->_cur_seg_size == this->segment_size(), "sanity"); |
aoqi@0 | 222 | } |
aoqi@0 | 223 | } |
aoqi@0 | 224 | |
aoqi@0 | 225 | template <class E, MEMFLAGS F> |
aoqi@0 | 226 | void Stack<E, F>::zap_segment(E* seg, bool zap_link_field) const |
aoqi@0 | 227 | { |
aoqi@0 | 228 | if (!ZapStackSegments) return; |
aoqi@0 | 229 | const size_t zap_bytes = segment_bytes() - (zap_link_field ? 0 : sizeof(E*)); |
aoqi@0 | 230 | uint32_t* cur = (uint32_t*)seg; |
aoqi@0 | 231 | const uint32_t* end = cur + zap_bytes / sizeof(uint32_t); |
aoqi@0 | 232 | while (cur < end) { |
aoqi@0 | 233 | *cur++ = 0xfadfaded; |
aoqi@0 | 234 | } |
aoqi@0 | 235 | } |
aoqi@0 | 236 | #endif |
aoqi@0 | 237 | |
aoqi@0 | 238 | template <class E, MEMFLAGS F> |
aoqi@0 | 239 | E* ResourceStack<E, F>::alloc(size_t bytes) |
aoqi@0 | 240 | { |
aoqi@0 | 241 | return (E*) resource_allocate_bytes(bytes); |
aoqi@0 | 242 | } |
aoqi@0 | 243 | |
aoqi@0 | 244 | template <class E, MEMFLAGS F> |
aoqi@0 | 245 | void ResourceStack<E, F>::free(E* addr, size_t bytes) |
aoqi@0 | 246 | { |
aoqi@0 | 247 | resource_free_bytes((char*) addr, bytes); |
aoqi@0 | 248 | } |
aoqi@0 | 249 | |
aoqi@0 | 250 | template <class E, MEMFLAGS F> |
aoqi@0 | 251 | void StackIterator<E, F>::sync() |
aoqi@0 | 252 | { |
aoqi@0 | 253 | _full_seg_size = _stack._full_seg_size; |
aoqi@0 | 254 | _cur_seg_size = _stack._cur_seg_size; |
aoqi@0 | 255 | _cur_seg = _stack._cur_seg; |
aoqi@0 | 256 | } |
aoqi@0 | 257 | |
aoqi@0 | 258 | template <class E, MEMFLAGS F> |
aoqi@0 | 259 | E* StackIterator<E, F>::next_addr() |
aoqi@0 | 260 | { |
aoqi@0 | 261 | assert(!is_empty(), "no items left"); |
aoqi@0 | 262 | if (_cur_seg_size == 1) { |
aoqi@0 | 263 | E* addr = _cur_seg; |
aoqi@0 | 264 | _cur_seg = _stack.get_link(_cur_seg); |
aoqi@0 | 265 | _cur_seg_size = _stack.segment_size(); |
aoqi@0 | 266 | _full_seg_size -= _stack.segment_size(); |
aoqi@0 | 267 | return addr; |
aoqi@0 | 268 | } |
aoqi@0 | 269 | return _cur_seg + --_cur_seg_size; |
aoqi@0 | 270 | } |
aoqi@0 | 271 | |
aoqi@0 | 272 | #endif // SHARE_VM_UTILITIES_STACK_INLINE_HPP |