Thu, 22 Apr 2010 10:02:38 -0700
6819061: G1: eliminate serial Other times that are proportional to the collection set length
6871109: G1: remove the concept of the scan only prefix
Summary: Removed scan only regions and associated code. The young portion of the collection set is now constructed incrementally - when a young region is retired as the current allocation region it is added to the collection set.
Reviewed-by: apetrusenko, iveresov, tonyp
1 /*
2 * Copyright 2001-2009 Sun Microsystems, Inc. All Rights Reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
20 * CA 95054 USA or visit www.sun.com if you need additional information or
21 * have any questions.
22 *
23 */
25 // Forward decl
26 class ConcurrentG1RefineThread;
27 class G1RemSet;
29 class ConcurrentG1Refine: public CHeapObj {
30 ConcurrentG1RefineThread** _threads;
31 int _n_threads;
32 int _n_worker_threads;
33 /*
34 * The value of the update buffer queue length falls into one of 3 zones:
35 * green, yellow, red. If the value is in [0, green) nothing is
36 * done, the buffers are left unprocessed to enable the caching effect of the
37 * dirtied cards. In the yellow zone [green, yellow) the concurrent refinement
38 * threads are gradually activated. In [yellow, red) all threads are
39 * running. If the length becomes red (max queue length) the mutators start
40 * processing the buffers.
41 *
42 * There are some interesting cases (when G1UseAdaptiveConcRefinement
43 * is turned off):
44 * 1) green = yellow = red = 0. In this case the mutator will process all
45 * buffers. Except for those that are created by the deferred updates
46 * machinery during a collection.
47 * 2) green = 0. Means no caching. Can be a good way to minimize the
48 * amount of time spent updating rsets during a collection.
49 */
50 int _green_zone;
51 int _yellow_zone;
52 int _red_zone;
54 int _thread_threshold_step;
56 // Reset the threshold step value based of the current zone boundaries.
57 void reset_threshold_step();
59 // The cache for card refinement.
60 bool _use_cache;
61 bool _def_use_cache;
63 size_t _n_periods; // Used as clearing epoch
65 // An evicting cache of the number of times each card
66 // is accessed. Reduces, but does not eliminate, the amount
67 // of duplicated processing of dirty cards.
69 enum SomePrivateConstants {
70 epoch_bits = 32,
71 card_num_shift = epoch_bits,
72 epoch_mask = AllBits,
73 card_num_mask = AllBits,
75 // The initial cache size is approximately this fraction
76 // of a maximal cache (i.e. the size needed for all cards
77 // in the heap)
78 InitialCacheFraction = 512
79 };
81 const static julong card_num_mask_in_place =
82 (julong) card_num_mask << card_num_shift;
84 typedef struct {
85 julong _value; // | card_num | epoch |
86 } CardEpochCacheEntry;
88 julong make_epoch_entry(unsigned int card_num, unsigned int epoch) {
89 assert(0 <= card_num && card_num < _max_n_card_counts, "Bounds");
90 assert(0 <= epoch && epoch <= _n_periods, "must be");
92 return ((julong) card_num << card_num_shift) | epoch;
93 }
95 unsigned int extract_epoch(julong v) {
96 return (v & epoch_mask);
97 }
99 unsigned int extract_card_num(julong v) {
100 return (v & card_num_mask_in_place) >> card_num_shift;
101 }
103 typedef struct {
104 unsigned char _count;
105 unsigned char _evict_count;
106 } CardCountCacheEntry;
108 CardCountCacheEntry* _card_counts;
109 CardEpochCacheEntry* _card_epochs;
111 // The current number of buckets in the card count cache
112 unsigned _n_card_counts;
114 // The max number of buckets required for the number of
115 // cards for the entire reserved heap
116 unsigned _max_n_card_counts;
118 // Possible sizes of the cache: odd primes that roughly double in size.
119 // (See jvmtiTagMap.cpp).
120 static int _cc_cache_sizes[];
122 // The index in _cc_cache_sizes corresponding to the size of
123 // _card_counts.
124 int _cache_size_index;
126 bool _expand_card_counts;
128 const jbyte* _ct_bot;
130 jbyte** _hot_cache;
131 int _hot_cache_size;
132 int _n_hot;
133 int _hot_cache_idx;
135 int _hot_cache_par_chunk_size;
136 volatile int _hot_cache_par_claimed_idx;
138 // Needed to workaround 6817995
139 CardTableModRefBS* _ct_bs;
140 G1CollectedHeap* _g1h;
142 // Expands the array that holds the card counts to the next size up
143 void expand_card_count_cache();
145 // hash a given key (index of card_ptr) with the specified size
146 static unsigned int hash(size_t key, int size) {
147 return (unsigned int) key % size;
148 }
150 // hash a given key (index of card_ptr)
151 unsigned int hash(size_t key) {
152 return hash(key, _n_card_counts);
153 }
155 unsigned ptr_2_card_num(jbyte* card_ptr) {
156 return (unsigned) (card_ptr - _ct_bot);
157 }
159 jbyte* card_num_2_ptr(unsigned card_num) {
160 return (jbyte*) (_ct_bot + card_num);
161 }
163 // Returns the count of this card after incrementing it.
164 jbyte* add_card_count(jbyte* card_ptr, int* count, bool* defer);
166 // Returns true if this card is in a young region
167 bool is_young_card(jbyte* card_ptr);
169 public:
170 ConcurrentG1Refine();
171 ~ConcurrentG1Refine();
173 void init(); // Accomplish some initialization that has to wait.
174 void stop();
176 void reinitialize_threads();
178 // Iterate over the conc refine threads
179 void threads_do(ThreadClosure *tc);
181 // If this is the first entry for the slot, writes into the cache and
182 // returns NULL. If it causes an eviction, returns the evicted pointer.
183 // Otherwise, its a cache hit, and returns NULL.
184 jbyte* cache_insert(jbyte* card_ptr, bool* defer);
186 // Process the cached entries.
187 void clean_up_cache(int worker_i, G1RemSet* g1rs);
189 // Set up for parallel processing of the cards in the hot cache
190 void clear_hot_cache_claimed_index() {
191 _hot_cache_par_claimed_idx = 0;
192 }
194 // Discard entries in the hot cache.
195 void clear_hot_cache() {
196 _hot_cache_idx = 0; _n_hot = 0;
197 }
199 bool hot_cache_is_empty() { return _n_hot == 0; }
201 bool use_cache() { return _use_cache; }
202 void set_use_cache(bool b) {
203 if (b) _use_cache = _def_use_cache;
204 else _use_cache = false;
205 }
207 void clear_and_record_card_counts();
209 static int thread_num();
211 void print_worker_threads_on(outputStream* st) const;
213 void set_green_zone(int x) { _green_zone = x; }
214 void set_yellow_zone(int x) { _yellow_zone = x; }
215 void set_red_zone(int x) { _red_zone = x; }
217 int green_zone() const { return _green_zone; }
218 int yellow_zone() const { return _yellow_zone; }
219 int red_zone() const { return _red_zone; }
221 int total_thread_num() const { return _n_threads; }
222 int worker_thread_num() const { return _n_worker_threads; }
224 int thread_threshold_step() const { return _thread_threshold_step; }
225 };