Wed, 26 Jun 2013 16:58:37 +0200
8013590: NPG: Add a memory pool MXBean for Metaspace
Reviewed-by: jmasa, mgerdin
duke@435 | 1 | /* |
coleenp@4037 | 2 | * Copyright (c) 2000, 2012, Oracle and/or its affiliates. All rights reserved. |
duke@435 | 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
duke@435 | 4 | * |
duke@435 | 5 | * This code is free software; you can redistribute it and/or modify it |
duke@435 | 6 | * under the terms of the GNU General Public License version 2 only, as |
duke@435 | 7 | * published by the Free Software Foundation. |
duke@435 | 8 | * |
duke@435 | 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
duke@435 | 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
duke@435 | 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
duke@435 | 12 | * version 2 for more details (a copy is included in the LICENSE file that |
duke@435 | 13 | * accompanied this code). |
duke@435 | 14 | * |
duke@435 | 15 | * You should have received a copy of the GNU General Public License version |
duke@435 | 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
duke@435 | 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
duke@435 | 18 | * |
trims@1907 | 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
trims@1907 | 20 | * or visit www.oracle.com if you need additional information or have any |
trims@1907 | 21 | * questions. |
duke@435 | 22 | * |
duke@435 | 23 | */ |
duke@435 | 24 | |
stefank@2314 | 25 | #ifndef SHARE_VM_MEMORY_BARRIERSET_HPP |
stefank@2314 | 26 | #define SHARE_VM_MEMORY_BARRIERSET_HPP |
stefank@2314 | 27 | |
stefank@2314 | 28 | #include "memory/memRegion.hpp" |
stefank@2314 | 29 | #include "oops/oopsHierarchy.hpp" |
stefank@2314 | 30 | |
duke@435 | 31 | // This class provides the interface between a barrier implementation and |
duke@435 | 32 | // the rest of the system. |
duke@435 | 33 | |
zgu@3900 | 34 | class BarrierSet: public CHeapObj<mtGC> { |
duke@435 | 35 | friend class VMStructs; |
duke@435 | 36 | public: |
duke@435 | 37 | enum Name { |
duke@435 | 38 | ModRef, |
duke@435 | 39 | CardTableModRef, |
duke@435 | 40 | CardTableExtension, |
ysr@777 | 41 | G1SATBCT, |
ysr@777 | 42 | G1SATBCTLogging, |
duke@435 | 43 | Other, |
duke@435 | 44 | Uninit |
duke@435 | 45 | }; |
duke@435 | 46 | |
iveresov@2606 | 47 | enum Flags { |
iveresov@2606 | 48 | None = 0, |
iveresov@2606 | 49 | TargetUninitialized = 1 |
iveresov@2606 | 50 | }; |
duke@435 | 51 | protected: |
duke@435 | 52 | int _max_covered_regions; |
duke@435 | 53 | Name _kind; |
duke@435 | 54 | |
duke@435 | 55 | public: |
duke@435 | 56 | |
ysr@777 | 57 | BarrierSet() { _kind = Uninit; } |
duke@435 | 58 | // To get around prohibition on RTTI. |
ysr@777 | 59 | BarrierSet::Name kind() { return _kind; } |
duke@435 | 60 | virtual bool is_a(BarrierSet::Name bsn) = 0; |
duke@435 | 61 | |
duke@435 | 62 | // These operations indicate what kind of barriers the BarrierSet has. |
duke@435 | 63 | virtual bool has_read_ref_barrier() = 0; |
duke@435 | 64 | virtual bool has_read_prim_barrier() = 0; |
duke@435 | 65 | virtual bool has_write_ref_barrier() = 0; |
ysr@777 | 66 | virtual bool has_write_ref_pre_barrier() = 0; |
duke@435 | 67 | virtual bool has_write_prim_barrier() = 0; |
duke@435 | 68 | |
duke@435 | 69 | // These functions indicate whether a particular access of the given |
duke@435 | 70 | // kinds requires a barrier. |
coleenp@548 | 71 | virtual bool read_ref_needs_barrier(void* field) = 0; |
duke@435 | 72 | virtual bool read_prim_needs_barrier(HeapWord* field, size_t bytes) = 0; |
ysr@777 | 73 | virtual bool write_prim_needs_barrier(HeapWord* field, size_t bytes, |
ysr@777 | 74 | juint val1, juint val2) = 0; |
duke@435 | 75 | |
duke@435 | 76 | // The first four operations provide a direct implementation of the |
duke@435 | 77 | // barrier set. An interpreter loop, for example, could call these |
duke@435 | 78 | // directly, as appropriate. |
duke@435 | 79 | |
duke@435 | 80 | // Invoke the barrier, if any, necessary when reading the given ref field. |
coleenp@548 | 81 | virtual void read_ref_field(void* field) = 0; |
duke@435 | 82 | |
duke@435 | 83 | // Invoke the barrier, if any, necessary when reading the given primitive |
duke@435 | 84 | // "field" of "bytes" bytes in "obj". |
duke@435 | 85 | virtual void read_prim_field(HeapWord* field, size_t bytes) = 0; |
duke@435 | 86 | |
duke@435 | 87 | // Invoke the barrier, if any, necessary when writing "new_val" into the |
duke@435 | 88 | // ref field at "offset" in "obj". |
duke@435 | 89 | // (For efficiency reasons, this operation is specialized for certain |
duke@435 | 90 | // barrier types. Semantically, it should be thought of as a call to the |
duke@435 | 91 | // virtual "_work" function below, which must implement the barrier.) |
ysr@777 | 92 | // First the pre-write versions... |
ysr@1280 | 93 | template <class T> inline void write_ref_field_pre(T* field, oop new_val); |
ysr@1280 | 94 | private: |
ysr@1280 | 95 | // Keep this private so as to catch violations at build time. |
ysr@1280 | 96 | virtual void write_ref_field_pre_work( void* field, oop new_val) { guarantee(false, "Not needed"); }; |
ysr@777 | 97 | protected: |
ysr@1280 | 98 | virtual void write_ref_field_pre_work( oop* field, oop new_val) {}; |
ysr@1280 | 99 | virtual void write_ref_field_pre_work(narrowOop* field, oop new_val) {}; |
ysr@777 | 100 | public: |
ysr@777 | 101 | |
ysr@777 | 102 | // ...then the post-write version. |
coleenp@548 | 103 | inline void write_ref_field(void* field, oop new_val); |
duke@435 | 104 | protected: |
coleenp@548 | 105 | virtual void write_ref_field_work(void* field, oop new_val) = 0; |
duke@435 | 106 | public: |
duke@435 | 107 | |
duke@435 | 108 | // Invoke the barrier, if any, necessary when writing the "bytes"-byte |
duke@435 | 109 | // value(s) "val1" (and "val2") into the primitive "field". |
duke@435 | 110 | virtual void write_prim_field(HeapWord* field, size_t bytes, |
duke@435 | 111 | juint val1, juint val2) = 0; |
duke@435 | 112 | |
duke@435 | 113 | // Operations on arrays, or general regions (e.g., for "clone") may be |
duke@435 | 114 | // optimized by some barriers. |
duke@435 | 115 | |
duke@435 | 116 | // The first six operations tell whether such an optimization exists for |
duke@435 | 117 | // the particular barrier. |
duke@435 | 118 | virtual bool has_read_ref_array_opt() = 0; |
duke@435 | 119 | virtual bool has_read_prim_array_opt() = 0; |
ysr@777 | 120 | virtual bool has_write_ref_array_pre_opt() { return true; } |
duke@435 | 121 | virtual bool has_write_ref_array_opt() = 0; |
duke@435 | 122 | virtual bool has_write_prim_array_opt() = 0; |
duke@435 | 123 | |
duke@435 | 124 | virtual bool has_read_region_opt() = 0; |
duke@435 | 125 | virtual bool has_write_region_opt() = 0; |
duke@435 | 126 | |
duke@435 | 127 | // These operations should assert false unless the correponding operation |
duke@435 | 128 | // above returns true. Otherwise, they should perform an appropriate |
duke@435 | 129 | // barrier for an array whose elements are all in the given memory region. |
duke@435 | 130 | virtual void read_ref_array(MemRegion mr) = 0; |
duke@435 | 131 | virtual void read_prim_array(MemRegion mr) = 0; |
duke@435 | 132 | |
ysr@1526 | 133 | // Below length is the # array elements being written |
iveresov@2606 | 134 | virtual void write_ref_array_pre(oop* dst, int length, |
iveresov@2606 | 135 | bool dest_uninitialized = false) {} |
iveresov@2606 | 136 | virtual void write_ref_array_pre(narrowOop* dst, int length, |
iveresov@2606 | 137 | bool dest_uninitialized = false) {} |
ysr@1526 | 138 | // Below count is the # array elements being written, starting |
ysr@1526 | 139 | // at the address "start", which may not necessarily be HeapWord-aligned |
ysr@1526 | 140 | inline void write_ref_array(HeapWord* start, size_t count); |
ysr@777 | 141 | |
ysr@1526 | 142 | // Static versions, suitable for calling from generated code; |
ysr@1526 | 143 | // count is # array elements being written, starting with "start", |
ysr@1526 | 144 | // which may not necessarily be HeapWord-aligned. |
ysr@777 | 145 | static void static_write_ref_array_pre(HeapWord* start, size_t count); |
ysr@777 | 146 | static void static_write_ref_array_post(HeapWord* start, size_t count); |
ysr@777 | 147 | |
duke@435 | 148 | protected: |
duke@435 | 149 | virtual void write_ref_array_work(MemRegion mr) = 0; |
duke@435 | 150 | public: |
duke@435 | 151 | virtual void write_prim_array(MemRegion mr) = 0; |
duke@435 | 152 | |
duke@435 | 153 | virtual void read_region(MemRegion mr) = 0; |
duke@435 | 154 | |
duke@435 | 155 | // (For efficiency reasons, this operation is specialized for certain |
duke@435 | 156 | // barrier types. Semantically, it should be thought of as a call to the |
duke@435 | 157 | // virtual "_work" function below, which must implement the barrier.) |
duke@435 | 158 | inline void write_region(MemRegion mr); |
duke@435 | 159 | protected: |
duke@435 | 160 | virtual void write_region_work(MemRegion mr) = 0; |
duke@435 | 161 | public: |
duke@435 | 162 | |
duke@435 | 163 | // Some barrier sets create tables whose elements correspond to parts of |
duke@435 | 164 | // the heap; the CardTableModRefBS is an example. Such barrier sets will |
duke@435 | 165 | // normally reserve space for such tables, and commit parts of the table |
duke@435 | 166 | // "covering" parts of the heap that are committed. The constructor is |
duke@435 | 167 | // passed the maximum number of independently committable subregions to |
duke@435 | 168 | // be covered, and the "resize_covoered_region" function allows the |
duke@435 | 169 | // sub-parts of the heap to inform the barrier set of changes of their |
duke@435 | 170 | // sizes. |
duke@435 | 171 | BarrierSet(int max_covered_regions) : |
duke@435 | 172 | _max_covered_regions(max_covered_regions) {} |
duke@435 | 173 | |
duke@435 | 174 | // Inform the BarrierSet that the the covered heap region that starts |
duke@435 | 175 | // with "base" has been changed to have the given size (possibly from 0, |
duke@435 | 176 | // for initialization.) |
duke@435 | 177 | virtual void resize_covered_region(MemRegion new_region) = 0; |
duke@435 | 178 | |
duke@435 | 179 | // If the barrier set imposes any alignment restrictions on boundaries |
duke@435 | 180 | // within the heap, this function tells whether they are met. |
duke@435 | 181 | virtual bool is_aligned(HeapWord* addr) = 0; |
duke@435 | 182 | |
never@3687 | 183 | // Print a description of the memory for the barrier set |
never@3687 | 184 | virtual void print_on(outputStream* st) const = 0; |
duke@435 | 185 | }; |
stefank@2314 | 186 | |
stefank@2314 | 187 | #endif // SHARE_VM_MEMORY_BARRIERSET_HPP |