Wed, 15 Feb 2012 16:29:40 -0800
7142680: default GC affected by jvm path
Summary: Removed old tiered code
Reviewed-by: never, kvn
1 /*
2 * Copyright (c) 1997, 2011, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
25 #ifndef SHARE_VM_RUNTIME_VM_OPERATIONS_HPP
26 #define SHARE_VM_RUNTIME_VM_OPERATIONS_HPP
28 #include "classfile/javaClasses.hpp"
29 #include "memory/allocation.hpp"
30 #include "oops/oop.hpp"
31 #include "runtime/thread.hpp"
32 #include "utilities/top.hpp"
34 // The following classes are used for operations
35 // initiated by a Java thread but that must
36 // take place in the VMThread.
38 #define VM_OP_ENUM(type) VMOp_##type,
40 // Note: When new VM_XXX comes up, add 'XXX' to the template table.
41 #define VM_OPS_DO(template) \
42 template(Dummy) \
43 template(ThreadStop) \
44 template(ThreadDump) \
45 template(PrintThreads) \
46 template(FindDeadlocks) \
47 template(ForceSafepoint) \
48 template(ForceAsyncSafepoint) \
49 template(Deoptimize) \
50 template(DeoptimizeFrame) \
51 template(DeoptimizeAll) \
52 template(ZombieAll) \
53 template(UnlinkSymbols) \
54 template(HandleFullCodeCache) \
55 template(Verify) \
56 template(PrintJNI) \
57 template(HeapDumper) \
58 template(DeoptimizeTheWorld) \
59 template(GC_HeapInspection) \
60 template(GenCollectFull) \
61 template(GenCollectFullConcurrent) \
62 template(GenCollectForAllocation) \
63 template(GenCollectForPermanentAllocation) \
64 template(ParallelGCFailedAllocation) \
65 template(ParallelGCFailedPermanentAllocation) \
66 template(ParallelGCSystemGC) \
67 template(CGC_Operation) \
68 template(CMS_Initial_Mark) \
69 template(CMS_Final_Remark) \
70 template(G1CollectFull) \
71 template(G1CollectForAllocation) \
72 template(G1IncCollectionPause) \
73 template(EnableBiasedLocking) \
74 template(RevokeBias) \
75 template(BulkRevokeBias) \
76 template(PopulateDumpSharedSpace) \
77 template(JNIFunctionTableCopier) \
78 template(RedefineClasses) \
79 template(GetOwnedMonitorInfo) \
80 template(GetObjectMonitorUsage) \
81 template(GetCurrentContendedMonitor) \
82 template(GetStackTrace) \
83 template(GetMultipleStackTraces) \
84 template(GetAllStackTraces) \
85 template(GetThreadListStackTraces) \
86 template(GetFrameCount) \
87 template(GetFrameLocation) \
88 template(ChangeBreakpoints) \
89 template(GetOrSetLocal) \
90 template(GetCurrentLocation) \
91 template(EnterInterpOnlyMode) \
92 template(ChangeSingleStep) \
93 template(HeapWalkOperation) \
94 template(HeapIterateOperation) \
95 template(ReportJavaOutOfMemory) \
96 template(JFRCheckpoint) \
97 template(Exit) \
99 class VM_Operation: public CHeapObj {
100 public:
101 enum Mode {
102 _safepoint, // blocking, safepoint, vm_op C-heap allocated
103 _no_safepoint, // blocking, no safepoint, vm_op C-Heap allocated
104 _concurrent, // non-blocking, no safepoint, vm_op C-Heap allocated
105 _async_safepoint // non-blocking, safepoint, vm_op C-Heap allocated
106 };
108 enum VMOp_Type {
109 VM_OPS_DO(VM_OP_ENUM)
110 VMOp_Terminating
111 };
113 private:
114 Thread* _calling_thread;
115 ThreadPriority _priority;
116 long _timestamp;
117 VM_Operation* _next;
118 VM_Operation* _prev;
120 // The VM operation name array
121 static const char* _names[];
123 public:
124 VM_Operation() { _calling_thread = NULL; _next = NULL; _prev = NULL; }
125 virtual ~VM_Operation() {}
127 // VM operation support (used by VM thread)
128 Thread* calling_thread() const { return _calling_thread; }
129 ThreadPriority priority() { return _priority; }
130 void set_calling_thread(Thread* thread, ThreadPriority priority);
132 long timestamp() const { return _timestamp; }
133 void set_timestamp(long timestamp) { _timestamp = timestamp; }
135 // Called by VM thread - does in turn invoke doit(). Do not override this
136 void evaluate();
138 // evaluate() is called by the VMThread and in turn calls doit().
139 // If the thread invoking VMThread::execute((VM_Operation*) is a JavaThread,
140 // doit_prologue() is called in that thread before transferring control to
141 // the VMThread.
142 // If doit_prologue() returns true the VM operation will proceed, and
143 // doit_epilogue() will be called by the JavaThread once the VM operation
144 // completes. If doit_prologue() returns false the VM operation is cancelled.
145 virtual void doit() = 0;
146 virtual bool doit_prologue() { return true; };
147 virtual void doit_epilogue() {}; // Note: Not called if mode is: _concurrent
149 // Type test
150 virtual bool is_methodCompiler() const { return false; }
152 // Linking
153 VM_Operation *next() const { return _next; }
154 VM_Operation *prev() const { return _prev; }
155 void set_next(VM_Operation *next) { _next = next; }
156 void set_prev(VM_Operation *prev) { _prev = prev; }
158 // Configuration. Override these appropriatly in subclasses.
159 virtual VMOp_Type type() const = 0;
160 virtual Mode evaluation_mode() const { return _safepoint; }
161 virtual bool allow_nested_vm_operations() const { return false; }
162 virtual bool is_cheap_allocated() const { return false; }
163 virtual void oops_do(OopClosure* f) { /* do nothing */ };
165 // CAUTION: <don't hang yourself with following rope>
166 // If you override these methods, make sure that the evaluation
167 // of these methods is race-free and non-blocking, since these
168 // methods may be evaluated either by the mutators or by the
169 // vm thread, either concurrently with mutators or with the mutators
170 // stopped. In other words, taking locks is verboten, and if there
171 // are any races in evaluating the conditions, they'd better be benign.
172 virtual bool evaluate_at_safepoint() const {
173 return evaluation_mode() == _safepoint ||
174 evaluation_mode() == _async_safepoint;
175 }
176 virtual bool evaluate_concurrently() const {
177 return evaluation_mode() == _concurrent ||
178 evaluation_mode() == _async_safepoint;
179 }
181 // Debugging
182 void print_on_error(outputStream* st) const;
183 const char* name() const { return _names[type()]; }
184 static const char* name(int type) {
185 assert(type >= 0 && type < VMOp_Terminating, "invalid VM operation type");
186 return _names[type];
187 }
188 #ifndef PRODUCT
189 void print_on(outputStream* st) const { print_on_error(st); }
190 #endif
191 };
193 class VM_ThreadStop: public VM_Operation {
194 private:
195 oop _thread; // The Thread that the Throwable is thrown against
196 oop _throwable; // The Throwable thrown at the target Thread
197 public:
198 // All oops are passed as JNI handles, since there is no guarantee that a GC might happen before the
199 // VM operation is executed.
200 VM_ThreadStop(oop thread, oop throwable) {
201 _thread = thread;
202 _throwable = throwable;
203 }
204 VMOp_Type type() const { return VMOp_ThreadStop; }
205 oop target_thread() const { return _thread; }
206 oop throwable() const { return _throwable;}
207 void doit();
208 // We deoptimize if top-most frame is compiled - this might require a C2I adapter to be generated
209 bool allow_nested_vm_operations() const { return true; }
210 Mode evaluation_mode() const { return _async_safepoint; }
211 bool is_cheap_allocated() const { return true; }
213 // GC support
214 void oops_do(OopClosure* f) {
215 f->do_oop(&_thread); f->do_oop(&_throwable);
216 }
217 };
219 // dummy vm op, evaluated just to force a safepoint
220 class VM_ForceSafepoint: public VM_Operation {
221 public:
222 VM_ForceSafepoint() {}
223 void doit() {}
224 VMOp_Type type() const { return VMOp_ForceSafepoint; }
225 };
227 // dummy vm op, evaluated just to force a safepoint
228 class VM_ForceAsyncSafepoint: public VM_Operation {
229 public:
230 VM_ForceAsyncSafepoint() {}
231 void doit() {}
232 VMOp_Type type() const { return VMOp_ForceAsyncSafepoint; }
233 Mode evaluation_mode() const { return _async_safepoint; }
234 bool is_cheap_allocated() const { return true; }
235 };
237 class VM_Deoptimize: public VM_Operation {
238 public:
239 VM_Deoptimize() {}
240 VMOp_Type type() const { return VMOp_Deoptimize; }
241 void doit();
242 bool allow_nested_vm_operations() const { return true; }
243 };
246 // Deopt helper that can deoptimize frames in threads other than the
247 // current thread. Only used through Deoptimization::deoptimize_frame.
248 class VM_DeoptimizeFrame: public VM_Operation {
249 friend class Deoptimization;
251 private:
252 JavaThread* _thread;
253 intptr_t* _id;
254 VM_DeoptimizeFrame(JavaThread* thread, intptr_t* id);
256 public:
257 VMOp_Type type() const { return VMOp_DeoptimizeFrame; }
258 void doit();
259 bool allow_nested_vm_operations() const { return true; }
260 };
262 class VM_HandleFullCodeCache: public VM_Operation {
263 private:
264 bool _is_full;
265 public:
266 VM_HandleFullCodeCache(bool is_full) { _is_full = is_full; }
267 VMOp_Type type() const { return VMOp_HandleFullCodeCache; }
268 void doit();
269 bool allow_nested_vm_operations() const { return true; }
270 };
272 #ifndef PRODUCT
273 class VM_DeoptimizeAll: public VM_Operation {
274 private:
275 KlassHandle _dependee;
276 public:
277 VM_DeoptimizeAll() {}
278 VMOp_Type type() const { return VMOp_DeoptimizeAll; }
279 void doit();
280 bool allow_nested_vm_operations() const { return true; }
281 };
284 class VM_ZombieAll: public VM_Operation {
285 public:
286 VM_ZombieAll() {}
287 VMOp_Type type() const { return VMOp_ZombieAll; }
288 void doit();
289 bool allow_nested_vm_operations() const { return true; }
290 };
291 #endif // PRODUCT
293 class VM_UnlinkSymbols: public VM_Operation {
294 public:
295 VM_UnlinkSymbols() {}
296 VMOp_Type type() const { return VMOp_UnlinkSymbols; }
297 void doit();
298 bool allow_nested_vm_operations() const { return true; }
299 };
301 class VM_Verify: public VM_Operation {
302 private:
303 KlassHandle _dependee;
304 public:
305 VM_Verify() {}
306 VMOp_Type type() const { return VMOp_Verify; }
307 void doit();
308 };
311 class VM_PrintThreads: public VM_Operation {
312 private:
313 outputStream* _out;
314 bool _print_concurrent_locks;
315 public:
316 VM_PrintThreads() { _out = tty; _print_concurrent_locks = PrintConcurrentLocks; }
317 VM_PrintThreads(outputStream* out, bool print_concurrent_locks) { _out = out; _print_concurrent_locks = print_concurrent_locks; }
318 VMOp_Type type() const { return VMOp_PrintThreads; }
319 void doit();
320 bool doit_prologue();
321 void doit_epilogue();
322 };
324 class VM_PrintJNI: public VM_Operation {
325 private:
326 outputStream* _out;
327 public:
328 VM_PrintJNI() { _out = tty; }
329 VM_PrintJNI(outputStream* out) { _out = out; }
330 VMOp_Type type() const { return VMOp_PrintJNI; }
331 void doit();
332 };
334 class DeadlockCycle;
335 class VM_FindDeadlocks: public VM_Operation {
336 private:
337 bool _concurrent_locks;
338 DeadlockCycle* _deadlocks;
339 outputStream* _out;
341 public:
342 VM_FindDeadlocks(bool concurrent_locks) : _concurrent_locks(concurrent_locks), _out(NULL), _deadlocks(NULL) {};
343 VM_FindDeadlocks(outputStream* st) : _concurrent_locks(true), _out(st), _deadlocks(NULL) {};
344 ~VM_FindDeadlocks();
346 DeadlockCycle* result() { return _deadlocks; };
347 VMOp_Type type() const { return VMOp_FindDeadlocks; }
348 void doit();
349 bool doit_prologue();
350 };
352 class ThreadDumpResult;
353 class ThreadSnapshot;
354 class ThreadConcurrentLocks;
356 class VM_ThreadDump : public VM_Operation {
357 private:
358 ThreadDumpResult* _result;
359 int _num_threads;
360 GrowableArray<instanceHandle>* _threads;
361 int _max_depth;
362 bool _with_locked_monitors;
363 bool _with_locked_synchronizers;
365 ThreadSnapshot* snapshot_thread(JavaThread* java_thread, ThreadConcurrentLocks* tcl);
367 public:
368 VM_ThreadDump(ThreadDumpResult* result,
369 int max_depth, // -1 indicates entire stack
370 bool with_locked_monitors,
371 bool with_locked_synchronizers);
373 VM_ThreadDump(ThreadDumpResult* result,
374 GrowableArray<instanceHandle>* threads,
375 int num_threads, // -1 indicates entire stack
376 int max_depth,
377 bool with_locked_monitors,
378 bool with_locked_synchronizers);
380 VMOp_Type type() const { return VMOp_ThreadDump; }
381 void doit();
382 bool doit_prologue();
383 void doit_epilogue();
384 };
387 class VM_Exit: public VM_Operation {
388 private:
389 int _exit_code;
390 static volatile bool _vm_exited;
391 static Thread * _shutdown_thread;
392 static void wait_if_vm_exited();
393 public:
394 VM_Exit(int exit_code) {
395 _exit_code = exit_code;
396 }
397 static int wait_for_threads_in_native_to_block();
398 static int set_vm_exited();
399 static bool vm_exited() { return _vm_exited; }
400 static void block_if_vm_exited() {
401 if (_vm_exited) {
402 wait_if_vm_exited();
403 }
404 }
405 VMOp_Type type() const { return VMOp_Exit; }
406 void doit();
407 };
409 #endif // SHARE_VM_RUNTIME_VM_OPERATIONS_HPP