Fri, 31 Oct 2014 12:16:20 +0100
8062169: Multiple OSR compilations issued for same bci
Summary: Fixed 'SimpleThresholdPolicy::event' to always perform OSR if an OSR nmethod is available.
Reviewed-by: kvn, iveresov
duke@435 | 1 | /* |
hseigel@6755 | 2 | * Copyright (c) 1997, 2014, Oracle and/or its affiliates. All rights reserved. |
duke@435 | 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
duke@435 | 4 | * |
duke@435 | 5 | * This code is free software; you can redistribute it and/or modify it |
duke@435 | 6 | * under the terms of the GNU General Public License version 2 only, as |
duke@435 | 7 | * published by the Free Software Foundation. |
duke@435 | 8 | * |
duke@435 | 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
duke@435 | 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
duke@435 | 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
duke@435 | 12 | * version 2 for more details (a copy is included in the LICENSE file that |
duke@435 | 13 | * accompanied this code). |
duke@435 | 14 | * |
duke@435 | 15 | * You should have received a copy of the GNU General Public License version |
duke@435 | 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
duke@435 | 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
duke@435 | 18 | * |
trims@1907 | 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
trims@1907 | 20 | * or visit www.oracle.com if you need additional information or have any |
trims@1907 | 21 | * questions. |
duke@435 | 22 | * |
duke@435 | 23 | */ |
duke@435 | 24 | |
stefank@2314 | 25 | #ifndef SHARE_VM_RUNTIME_OS_HPP |
stefank@2314 | 26 | #define SHARE_VM_RUNTIME_OS_HPP |
stefank@2314 | 27 | |
stefank@2314 | 28 | #include "jvmtifiles/jvmti.h" |
stefank@2314 | 29 | #include "runtime/atomic.hpp" |
stefank@2314 | 30 | #include "runtime/extendedPC.hpp" |
stefank@2314 | 31 | #include "runtime/handles.hpp" |
stefank@2314 | 32 | #include "utilities/top.hpp" |
stefank@2314 | 33 | #ifdef TARGET_OS_FAMILY_linux |
stefank@2314 | 34 | # include "jvm_linux.h" |
rbackman@5424 | 35 | # include <setjmp.h> |
stefank@2314 | 36 | #endif |
stefank@2314 | 37 | #ifdef TARGET_OS_FAMILY_solaris |
stefank@2314 | 38 | # include "jvm_solaris.h" |
rbackman@5424 | 39 | # include <setjmp.h> |
stefank@2314 | 40 | #endif |
stefank@2314 | 41 | #ifdef TARGET_OS_FAMILY_windows |
stefank@2314 | 42 | # include "jvm_windows.h" |
stefank@2314 | 43 | #endif |
goetz@6461 | 44 | #ifdef TARGET_OS_FAMILY_aix |
goetz@6461 | 45 | # include "jvm_aix.h" |
goetz@6461 | 46 | # include <setjmp.h> |
goetz@6461 | 47 | #endif |
never@3156 | 48 | #ifdef TARGET_OS_FAMILY_bsd |
never@3156 | 49 | # include "jvm_bsd.h" |
rbackman@5424 | 50 | # include <setjmp.h> |
sla@6667 | 51 | # ifdef __APPLE__ |
sla@6667 | 52 | # include <mach/mach_time.h> |
sla@6667 | 53 | # endif |
never@3156 | 54 | #endif |
stefank@2314 | 55 | |
bpittore@5585 | 56 | class AgentLibrary; |
bpittore@5585 | 57 | |
duke@435 | 58 | // os defines the interface to operating system; this includes traditional |
duke@435 | 59 | // OS services (time, I/O) as well as other functionality with system- |
duke@435 | 60 | // dependent code. |
duke@435 | 61 | |
duke@435 | 62 | typedef void (*dll_func)(...); |
duke@435 | 63 | |
duke@435 | 64 | class Thread; |
duke@435 | 65 | class JavaThread; |
duke@435 | 66 | class Event; |
duke@435 | 67 | class DLL; |
duke@435 | 68 | class FileHandle; |
zgu@7074 | 69 | class NativeCallStack; |
zgu@7074 | 70 | |
iveresov@576 | 71 | template<class E> class GrowableArray; |
duke@435 | 72 | |
duke@435 | 73 | // %%%%% Moved ThreadState, START_FN, OSThread to new osThread.hpp. -- Rose |
duke@435 | 74 | |
duke@435 | 75 | // Platform-independent error return values from OS functions |
duke@435 | 76 | enum OSReturn { |
duke@435 | 77 | OS_OK = 0, // Operation was successful |
duke@435 | 78 | OS_ERR = -1, // Operation failed |
duke@435 | 79 | OS_INTRPT = -2, // Operation was interrupted |
duke@435 | 80 | OS_TIMEOUT = -3, // Operation timed out |
duke@435 | 81 | OS_NOMEM = -5, // Operation failed for lack of memory |
duke@435 | 82 | OS_NORESOURCE = -6 // Operation failed for lack of nonmemory resource |
duke@435 | 83 | }; |
duke@435 | 84 | |
duke@435 | 85 | enum ThreadPriority { // JLS 20.20.1-3 |
duke@435 | 86 | NoPriority = -1, // Initial non-priority value |
duke@435 | 87 | MinPriority = 1, // Minimum priority |
duke@435 | 88 | NormPriority = 5, // Normal (non-daemon) priority |
duke@435 | 89 | NearMaxPriority = 9, // High priority, used for VMThread |
phh@3481 | 90 | MaxPriority = 10, // Highest priority, used for WatcherThread |
duke@435 | 91 | // ensures that VMThread doesn't starve profiler |
phh@3481 | 92 | CriticalPriority = 11 // Critical thread priority |
duke@435 | 93 | }; |
duke@435 | 94 | |
dcubed@5255 | 95 | // Executable parameter flag for os::commit_memory() and |
dcubed@5255 | 96 | // os::commit_memory_or_exit(). |
dcubed@5255 | 97 | const bool ExecMem = true; |
dcubed@5255 | 98 | |
duke@435 | 99 | // Typedef for structured exception handling support |
duke@435 | 100 | typedef void (*java_call_t)(JavaValue* value, methodHandle* method, JavaCallArguments* args, Thread* thread); |
duke@435 | 101 | |
zgu@7074 | 102 | class MallocTracker; |
zgu@7074 | 103 | |
duke@435 | 104 | class os: AllStatic { |
twisti@5726 | 105 | friend class VMStructs; |
zgu@7074 | 106 | friend class MallocTracker; |
phh@1558 | 107 | public: |
duke@435 | 108 | enum { page_sizes_max = 9 }; // Size of _page_sizes array (8 plus a sentinel) |
duke@435 | 109 | |
phh@1558 | 110 | private: |
duke@435 | 111 | static OSThread* _starting_thread; |
duke@435 | 112 | static address _polling_page; |
duke@435 | 113 | static volatile int32_t * _mem_serialize_page; |
duke@435 | 114 | static uintptr_t _serialize_page_mask; |
phh@1558 | 115 | public: |
duke@435 | 116 | static size_t _page_sizes[page_sizes_max]; |
duke@435 | 117 | |
phh@1558 | 118 | private: |
duke@435 | 119 | static void init_page_sizes(size_t default_page_size) { |
duke@435 | 120 | _page_sizes[0] = default_page_size; |
duke@435 | 121 | _page_sizes[1] = 0; // sentinel |
duke@435 | 122 | } |
duke@435 | 123 | |
zgu@3900 | 124 | static char* pd_reserve_memory(size_t bytes, char* addr = 0, |
zgu@3900 | 125 | size_t alignment_hint = 0); |
zgu@3900 | 126 | static char* pd_attempt_reserve_memory_at(size_t bytes, char* addr); |
zgu@3900 | 127 | static void pd_split_reserved_memory(char *base, size_t size, |
zgu@3900 | 128 | size_t split, bool realloc); |
dcubed@5255 | 129 | static bool pd_commit_memory(char* addr, size_t bytes, bool executable); |
zgu@3900 | 130 | static bool pd_commit_memory(char* addr, size_t size, size_t alignment_hint, |
dcubed@5255 | 131 | bool executable); |
dcubed@5255 | 132 | // Same as pd_commit_memory() that either succeeds or calls |
dcubed@5255 | 133 | // vm_exit_out_of_memory() with the specified mesg. |
dcubed@5255 | 134 | static void pd_commit_memory_or_exit(char* addr, size_t bytes, |
dcubed@5255 | 135 | bool executable, const char* mesg); |
dcubed@5255 | 136 | static void pd_commit_memory_or_exit(char* addr, size_t size, |
dcubed@5255 | 137 | size_t alignment_hint, |
dcubed@5255 | 138 | bool executable, const char* mesg); |
zgu@3900 | 139 | static bool pd_uncommit_memory(char* addr, size_t bytes); |
zgu@3900 | 140 | static bool pd_release_memory(char* addr, size_t bytes); |
zgu@3900 | 141 | |
zgu@3900 | 142 | static char* pd_map_memory(int fd, const char* file_name, size_t file_offset, |
zgu@3900 | 143 | char *addr, size_t bytes, bool read_only = false, |
zgu@3900 | 144 | bool allow_exec = false); |
zgu@3900 | 145 | static char* pd_remap_memory(int fd, const char* file_name, size_t file_offset, |
zgu@3900 | 146 | char *addr, size_t bytes, bool read_only, |
zgu@3900 | 147 | bool allow_exec); |
zgu@3900 | 148 | static bool pd_unmap_memory(char *addr, size_t bytes); |
zgu@3900 | 149 | static void pd_free_memory(char *addr, size_t bytes, size_t alignment_hint); |
zgu@3900 | 150 | static void pd_realign_memory(char *addr, size_t bytes, size_t alignment_hint); |
zgu@3900 | 151 | |
zgu@3900 | 152 | |
duke@435 | 153 | public: |
bobv@2036 | 154 | static void init(void); // Called before command line parsing |
tschatzl@5701 | 155 | static void init_before_ergo(void); // Called after command line parsing |
tschatzl@5701 | 156 | // before VM ergonomics processing. |
bobv@2036 | 157 | static jint init_2(void); // Called after command line parsing |
tschatzl@5701 | 158 | // and VM ergonomics processing |
phh@3378 | 159 | static void init_globals(void) { // Called from init_globals() in init.cpp |
phh@3378 | 160 | init_globals_ext(); |
phh@3378 | 161 | } |
bobv@2036 | 162 | static void init_3(void); // Called at the end of vm init |
duke@435 | 163 | |
duke@435 | 164 | // File names are case-insensitive on windows only |
duke@435 | 165 | // Override me as needed |
duke@435 | 166 | static int file_name_strcmp(const char* s1, const char* s2); |
duke@435 | 167 | |
zgu@7074 | 168 | // get/unset environment variable |
duke@435 | 169 | static bool getenv(const char* name, char* buffer, int len); |
zgu@7074 | 170 | static bool unsetenv(const char* name); |
zgu@7074 | 171 | |
duke@435 | 172 | static bool have_special_privileges(); |
duke@435 | 173 | |
duke@435 | 174 | static jlong javaTimeMillis(); |
duke@435 | 175 | static jlong javaTimeNanos(); |
duke@435 | 176 | static void javaTimeNanos_info(jvmtiTimerInfo *info_ptr); |
duke@435 | 177 | static void run_periodic_checks(); |
duke@435 | 178 | |
duke@435 | 179 | |
duke@435 | 180 | // Returns the elapsed time in seconds since the vm started. |
duke@435 | 181 | static double elapsedTime(); |
duke@435 | 182 | |
duke@435 | 183 | // Returns real time in seconds since an arbitrary point |
duke@435 | 184 | // in the past. |
duke@435 | 185 | static bool getTimesSecs(double* process_real_time, |
duke@435 | 186 | double* process_user_time, |
duke@435 | 187 | double* process_system_time); |
duke@435 | 188 | |
duke@435 | 189 | // Interface to the performance counter |
duke@435 | 190 | static jlong elapsed_counter(); |
duke@435 | 191 | static jlong elapsed_frequency(); |
duke@435 | 192 | |
ysr@777 | 193 | // The "virtual time" of a thread is the amount of time a thread has |
ysr@777 | 194 | // actually run. The first function indicates whether the OS supports |
ysr@777 | 195 | // this functionality for the current thread, and if so: |
ysr@777 | 196 | // * the second enables vtime tracking (if that is required). |
ysr@777 | 197 | // * the third tells whether vtime is enabled. |
ysr@777 | 198 | // * the fourth returns the elapsed virtual time for the current |
ysr@777 | 199 | // thread. |
ysr@777 | 200 | static bool supports_vtime(); |
ysr@777 | 201 | static bool enable_vtime(); |
ysr@777 | 202 | static bool vtime_enabled(); |
ysr@777 | 203 | static double elapsedVTime(); |
ysr@777 | 204 | |
duke@435 | 205 | // Return current local time in a string (YYYY-MM-DD HH:MM:SS). |
duke@435 | 206 | // It is MT safe, but not async-safe, as reading time zone |
duke@435 | 207 | // information may require a lock on some platforms. |
ysr@983 | 208 | static char* local_time_string(char *buf, size_t buflen); |
ysr@983 | 209 | static struct tm* localtime_pd (const time_t* clock, struct tm* res); |
duke@435 | 210 | // Fill in buffer with current local time as an ISO-8601 string. |
duke@435 | 211 | // E.g., YYYY-MM-DDThh:mm:ss.mmm+zzzz. |
duke@435 | 212 | // Returns buffer, or NULL if it failed. |
duke@435 | 213 | static char* iso8601_time(char* buffer, size_t buffer_length); |
duke@435 | 214 | |
duke@435 | 215 | // Interface for detecting multiprocessor system |
duke@435 | 216 | static inline bool is_MP() { |
dholmes@7273 | 217 | // During bootstrap if _processor_count is not yet initialized |
dholmes@7273 | 218 | // we claim to be MP as that is safest. If any platform has a |
dholmes@7273 | 219 | // stub generator that might be triggered in this phase and for |
dholmes@7273 | 220 | // which being declared MP when in fact not, is a problem - then |
dholmes@7273 | 221 | // the bootstrap routine for the stub generator needs to check |
dholmes@7273 | 222 | // the processor count directly and leave the bootstrap routine |
dholmes@7273 | 223 | // in place until called after initialization has ocurred. |
dholmes@7273 | 224 | return (_processor_count != 1) || AssumeMP; |
duke@435 | 225 | } |
duke@435 | 226 | static julong available_memory(); |
duke@435 | 227 | static julong physical_memory(); |
tschatzl@4854 | 228 | static bool has_allocatable_memory_limit(julong* limit); |
duke@435 | 229 | static bool is_server_class_machine(); |
duke@435 | 230 | |
duke@435 | 231 | // number of CPUs |
duke@435 | 232 | static int processor_count() { |
duke@435 | 233 | return _processor_count; |
duke@435 | 234 | } |
phh@1558 | 235 | static void set_processor_count(int count) { _processor_count = count; } |
duke@435 | 236 | |
duke@435 | 237 | // Returns the number of CPUs this process is currently allowed to run on. |
duke@435 | 238 | // Note that on some OSes this can change dynamically. |
duke@435 | 239 | static int active_processor_count(); |
duke@435 | 240 | |
duke@435 | 241 | // Bind processes to processors. |
duke@435 | 242 | // This is a two step procedure: |
duke@435 | 243 | // first you generate a distribution of processes to processors, |
duke@435 | 244 | // then you bind processes according to that distribution. |
duke@435 | 245 | // Compute a distribution for number of processes to processors. |
duke@435 | 246 | // Stores the processor id's into the distribution array argument. |
duke@435 | 247 | // Returns true if it worked, false if it didn't. |
duke@435 | 248 | static bool distribute_processes(uint length, uint* distribution); |
duke@435 | 249 | // Binds the current process to a processor. |
duke@435 | 250 | // Returns true if it worked, false if it didn't. |
duke@435 | 251 | static bool bind_to_processor(uint processor_id); |
duke@435 | 252 | |
dcubed@3202 | 253 | // Give a name to the current thread. |
dcubed@3202 | 254 | static void set_native_thread_name(const char *name); |
dcubed@3202 | 255 | |
duke@435 | 256 | // Interface for stack banging (predetect possible stack overflow for |
duke@435 | 257 | // exception processing) There are guard pages, and above that shadow |
duke@435 | 258 | // pages for stack overflow checking. |
duke@435 | 259 | static bool uses_stack_guard_pages(); |
duke@435 | 260 | static bool allocate_stack_guard_pages(); |
duke@435 | 261 | static void bang_stack_shadow_pages(); |
duke@435 | 262 | static bool stack_shadow_pages_available(Thread *thread, methodHandle method); |
duke@435 | 263 | |
duke@435 | 264 | // OS interface to Virtual Memory |
duke@435 | 265 | |
duke@435 | 266 | // Return the default page size. |
duke@435 | 267 | static int vm_page_size(); |
duke@435 | 268 | |
duke@435 | 269 | // Return the page size to use for a region of memory. The min_pages argument |
duke@435 | 270 | // is a hint intended to limit fragmentation; it says the returned page size |
duke@435 | 271 | // should be <= region_max_size / min_pages. Because min_pages is a hint, |
duke@435 | 272 | // this routine may return a size larger than region_max_size / min_pages. |
duke@435 | 273 | // |
duke@435 | 274 | // The current implementation ignores min_pages if a larger page size is an |
duke@435 | 275 | // exact multiple of both region_min_size and region_max_size. This allows |
duke@435 | 276 | // larger pages to be used when doing so would not cause fragmentation; in |
duke@435 | 277 | // particular, a single page can be used when region_min_size == |
duke@435 | 278 | // region_max_size == a supported page size. |
duke@435 | 279 | static size_t page_size_for_region(size_t region_min_size, |
duke@435 | 280 | size_t region_max_size, |
duke@435 | 281 | uint min_pages); |
tschatzl@5701 | 282 | // Return the largest page size that can be used |
tschatzl@5701 | 283 | static size_t max_page_size() { |
tschatzl@5701 | 284 | // The _page_sizes array is sorted in descending order. |
tschatzl@5701 | 285 | return _page_sizes[0]; |
tschatzl@5701 | 286 | } |
duke@435 | 287 | |
jcoomes@3057 | 288 | // Methods for tracing page sizes returned by the above method; enabled by |
duke@435 | 289 | // TracePageSizes. The region_{min,max}_size parameters should be the values |
duke@435 | 290 | // passed to page_size_for_region() and page_size should be the result of that |
duke@435 | 291 | // call. The (optional) base and size parameters should come from the |
duke@435 | 292 | // ReservedSpace base() and size() methods. |
jcoomes@3057 | 293 | static void trace_page_sizes(const char* str, const size_t* page_sizes, |
jcoomes@3057 | 294 | int count) PRODUCT_RETURN; |
duke@435 | 295 | static void trace_page_sizes(const char* str, const size_t region_min_size, |
duke@435 | 296 | const size_t region_max_size, |
duke@435 | 297 | const size_t page_size, |
duke@435 | 298 | const char* base = NULL, |
duke@435 | 299 | const size_t size = 0) PRODUCT_RETURN; |
duke@435 | 300 | |
duke@435 | 301 | static int vm_allocation_granularity(); |
duke@435 | 302 | static char* reserve_memory(size_t bytes, char* addr = 0, |
duke@435 | 303 | size_t alignment_hint = 0); |
zgu@5053 | 304 | static char* reserve_memory(size_t bytes, char* addr, |
zgu@5053 | 305 | size_t alignment_hint, MEMFLAGS flags); |
brutisso@4369 | 306 | static char* reserve_memory_aligned(size_t size, size_t alignment); |
duke@435 | 307 | static char* attempt_reserve_memory_at(size_t bytes, char* addr); |
duke@435 | 308 | static void split_reserved_memory(char *base, size_t size, |
duke@435 | 309 | size_t split, bool realloc); |
dcubed@5255 | 310 | static bool commit_memory(char* addr, size_t bytes, bool executable); |
coleenp@1091 | 311 | static bool commit_memory(char* addr, size_t size, size_t alignment_hint, |
dcubed@5255 | 312 | bool executable); |
dcubed@5255 | 313 | // Same as commit_memory() that either succeeds or calls |
dcubed@5255 | 314 | // vm_exit_out_of_memory() with the specified mesg. |
dcubed@5255 | 315 | static void commit_memory_or_exit(char* addr, size_t bytes, |
dcubed@5255 | 316 | bool executable, const char* mesg); |
dcubed@5255 | 317 | static void commit_memory_or_exit(char* addr, size_t size, |
dcubed@5255 | 318 | size_t alignment_hint, |
dcubed@5255 | 319 | bool executable, const char* mesg); |
duke@435 | 320 | static bool uncommit_memory(char* addr, size_t bytes); |
duke@435 | 321 | static bool release_memory(char* addr, size_t bytes); |
coleenp@672 | 322 | |
coleenp@672 | 323 | enum ProtType { MEM_PROT_NONE, MEM_PROT_READ, MEM_PROT_RW, MEM_PROT_RWX }; |
coleenp@672 | 324 | static bool protect_memory(char* addr, size_t bytes, ProtType prot, |
coleenp@912 | 325 | bool is_committed = true); |
coleenp@672 | 326 | |
duke@435 | 327 | static bool guard_memory(char* addr, size_t bytes); |
duke@435 | 328 | static bool unguard_memory(char* addr, size_t bytes); |
coleenp@1755 | 329 | static bool create_stack_guard_pages(char* addr, size_t bytes); |
zgu@3900 | 330 | static bool pd_create_stack_guard_pages(char* addr, size_t bytes); |
coleenp@1755 | 331 | static bool remove_stack_guard_pages(char* addr, size_t bytes); |
coleenp@1755 | 332 | |
duke@435 | 333 | static char* map_memory(int fd, const char* file_name, size_t file_offset, |
duke@435 | 334 | char *addr, size_t bytes, bool read_only = false, |
duke@435 | 335 | bool allow_exec = false); |
duke@435 | 336 | static char* remap_memory(int fd, const char* file_name, size_t file_offset, |
duke@435 | 337 | char *addr, size_t bytes, bool read_only, |
duke@435 | 338 | bool allow_exec); |
duke@435 | 339 | static bool unmap_memory(char *addr, size_t bytes); |
iveresov@3363 | 340 | static void free_memory(char *addr, size_t bytes, size_t alignment_hint); |
duke@435 | 341 | static void realign_memory(char *addr, size_t bytes, size_t alignment_hint); |
duke@435 | 342 | |
duke@435 | 343 | // NUMA-specific interface |
iveresov@576 | 344 | static bool numa_has_static_binding(); |
iveresov@576 | 345 | static bool numa_has_group_homing(); |
iveresov@576 | 346 | static void numa_make_local(char *addr, size_t bytes, int lgrp_hint); |
duke@435 | 347 | static void numa_make_global(char *addr, size_t bytes); |
duke@435 | 348 | static size_t numa_get_groups_num(); |
duke@435 | 349 | static size_t numa_get_leaf_groups(int *ids, size_t size); |
duke@435 | 350 | static bool numa_topology_changed(); |
duke@435 | 351 | static int numa_get_group_id(); |
duke@435 | 352 | |
duke@435 | 353 | // Page manipulation |
duke@435 | 354 | struct page_info { |
duke@435 | 355 | size_t size; |
duke@435 | 356 | int lgrp_id; |
duke@435 | 357 | }; |
duke@435 | 358 | static bool get_page_info(char *start, page_info* info); |
duke@435 | 359 | static char* scan_pages(char *start, char* end, page_info* page_expected, page_info* page_found); |
duke@435 | 360 | |
duke@435 | 361 | static char* non_memory_address_word(); |
duke@435 | 362 | // reserve, commit and pin the entire memory region |
stefank@5578 | 363 | static char* reserve_memory_special(size_t size, size_t alignment, |
stefank@5578 | 364 | char* addr, bool executable); |
duke@435 | 365 | static bool release_memory_special(char* addr, size_t bytes); |
iveresov@2850 | 366 | static void large_page_init(); |
duke@435 | 367 | static size_t large_page_size(); |
duke@435 | 368 | static bool can_commit_large_page_memory(); |
jcoomes@514 | 369 | static bool can_execute_large_page_memory(); |
duke@435 | 370 | |
duke@435 | 371 | // OS interface to polling page |
duke@435 | 372 | static address get_polling_page() { return _polling_page; } |
duke@435 | 373 | static void set_polling_page(address page) { _polling_page = page; } |
duke@435 | 374 | static bool is_poll_address(address addr) { return addr >= _polling_page && addr < (_polling_page + os::vm_page_size()); } |
duke@435 | 375 | static void make_polling_page_unreadable(); |
duke@435 | 376 | static void make_polling_page_readable(); |
duke@435 | 377 | |
duke@435 | 378 | // Routines used to serialize the thread state without using membars |
duke@435 | 379 | static void serialize_thread_states(); |
duke@435 | 380 | |
duke@435 | 381 | // Since we write to the serialize page from every thread, we |
duke@435 | 382 | // want stores to be on unique cache lines whenever possible |
duke@435 | 383 | // in order to minimize CPU cross talk. We pre-compute the |
duke@435 | 384 | // amount to shift the thread* to make this offset unique to |
duke@435 | 385 | // each thread. |
duke@435 | 386 | static int get_serialize_page_shift_count() { |
duke@435 | 387 | return SerializePageShiftCount; |
duke@435 | 388 | } |
duke@435 | 389 | |
duke@435 | 390 | static void set_serialize_page_mask(uintptr_t mask) { |
duke@435 | 391 | _serialize_page_mask = mask; |
duke@435 | 392 | } |
duke@435 | 393 | |
duke@435 | 394 | static unsigned int get_serialize_page_mask() { |
duke@435 | 395 | return _serialize_page_mask; |
duke@435 | 396 | } |
duke@435 | 397 | |
duke@435 | 398 | static void set_memory_serialize_page(address page); |
duke@435 | 399 | |
duke@435 | 400 | static address get_memory_serialize_page() { |
duke@435 | 401 | return (address)_mem_serialize_page; |
duke@435 | 402 | } |
duke@435 | 403 | |
duke@435 | 404 | static inline void write_memory_serialize_page(JavaThread *thread) { |
duke@435 | 405 | uintptr_t page_offset = ((uintptr_t)thread >> |
duke@435 | 406 | get_serialize_page_shift_count()) & |
duke@435 | 407 | get_serialize_page_mask(); |
duke@435 | 408 | *(volatile int32_t *)((uintptr_t)_mem_serialize_page+page_offset) = 1; |
duke@435 | 409 | } |
duke@435 | 410 | |
duke@435 | 411 | static bool is_memory_serialize_page(JavaThread *thread, address addr) { |
duke@435 | 412 | if (UseMembar) return false; |
twisti@1513 | 413 | // Previously this function calculated the exact address of this |
twisti@1513 | 414 | // thread's serialize page, and checked if the faulting address |
twisti@1513 | 415 | // was equal. However, some platforms mask off faulting addresses |
twisti@1513 | 416 | // to the page size, so now we just check that the address is |
twisti@1513 | 417 | // within the page. This makes the thread argument unnecessary, |
twisti@1513 | 418 | // but we retain the NULL check to preserve existing behaviour. |
duke@435 | 419 | if (thread == NULL) return false; |
twisti@1513 | 420 | address page = (address) _mem_serialize_page; |
twisti@1513 | 421 | return addr >= page && addr < (page + os::vm_page_size()); |
duke@435 | 422 | } |
duke@435 | 423 | |
duke@435 | 424 | static void block_on_serialize_page_trap(); |
duke@435 | 425 | |
duke@435 | 426 | // threads |
duke@435 | 427 | |
duke@435 | 428 | enum ThreadType { |
duke@435 | 429 | vm_thread, |
duke@435 | 430 | cgc_thread, // Concurrent GC thread |
duke@435 | 431 | pgc_thread, // Parallel GC thread |
duke@435 | 432 | java_thread, |
duke@435 | 433 | compiler_thread, |
bobv@2036 | 434 | watcher_thread, |
bobv@2036 | 435 | os_thread |
duke@435 | 436 | }; |
duke@435 | 437 | |
duke@435 | 438 | static bool create_thread(Thread* thread, |
duke@435 | 439 | ThreadType thr_type, |
duke@435 | 440 | size_t stack_size = 0); |
duke@435 | 441 | static bool create_main_thread(JavaThread* thread); |
duke@435 | 442 | static bool create_attached_thread(JavaThread* thread); |
duke@435 | 443 | static void pd_start_thread(Thread* thread); |
duke@435 | 444 | static void start_thread(Thread* thread); |
duke@435 | 445 | |
zgu@4079 | 446 | static void initialize_thread(Thread* thr); |
duke@435 | 447 | static void free_thread(OSThread* osthread); |
duke@435 | 448 | |
duke@435 | 449 | // thread id on Linux/64bit is 64bit, on Windows and Solaris, it's 32bit |
duke@435 | 450 | static intx current_thread_id(); |
duke@435 | 451 | static int current_process_id(); |
duke@435 | 452 | static int sleep(Thread* thread, jlong ms, bool interruptable); |
dsimms@6348 | 453 | // Short standalone OS sleep suitable for slow path spin loop. |
dsimms@6348 | 454 | // Ignores Thread.interrupt() (so keep it short). |
dsimms@6348 | 455 | // ms = 0, will sleep for the least amount of time allowed by the OS. |
dsimms@6348 | 456 | static void naked_short_sleep(jlong ms); |
duke@435 | 457 | static void infinite_sleep(); // never returns, use with CAUTION |
duke@435 | 458 | static void yield(); // Yields to all threads with same priority |
duke@435 | 459 | enum YieldResult { |
duke@435 | 460 | YIELD_SWITCHED = 1, // caller descheduled, other ready threads exist & ran |
duke@435 | 461 | YIELD_NONEREADY = 0, // No other runnable/ready threads. |
duke@435 | 462 | // platform-specific yield return immediately |
duke@435 | 463 | YIELD_UNKNOWN = -1 // Unknown: platform doesn't support _SWITCHED or _NONEREADY |
duke@435 | 464 | // YIELD_SWITCHED and YIELD_NONREADY imply the platform supports a "strong" |
duke@435 | 465 | // yield that can be used in lieu of blocking. |
duke@435 | 466 | } ; |
duke@435 | 467 | static YieldResult NakedYield () ; |
duke@435 | 468 | static void yield_all(int attempts = 0); // Yields to all other threads including lower priority |
duke@435 | 469 | static void loop_breaker(int attempts); // called from within tight loops to possibly influence time-sharing |
duke@435 | 470 | static OSReturn set_priority(Thread* thread, ThreadPriority priority); |
duke@435 | 471 | static OSReturn get_priority(const Thread* const thread, ThreadPriority& priority); |
duke@435 | 472 | |
duke@435 | 473 | static void interrupt(Thread* thread); |
duke@435 | 474 | static bool is_interrupted(Thread* thread, bool clear_interrupted); |
duke@435 | 475 | |
duke@435 | 476 | static int pd_self_suspend_thread(Thread* thread); |
duke@435 | 477 | |
duke@435 | 478 | static ExtendedPC fetch_frame_from_context(void* ucVoid, intptr_t** sp, intptr_t** fp); |
duke@435 | 479 | static frame fetch_frame_from_context(void* ucVoid); |
duke@435 | 480 | |
duke@435 | 481 | static ExtendedPC get_thread_pc(Thread *thread); |
duke@435 | 482 | static void breakpoint(); |
duke@435 | 483 | |
duke@435 | 484 | static address current_stack_pointer(); |
duke@435 | 485 | static address current_stack_base(); |
duke@435 | 486 | static size_t current_stack_size(); |
duke@435 | 487 | |
roland@3606 | 488 | static void verify_stack_alignment() PRODUCT_RETURN; |
roland@3606 | 489 | |
duke@435 | 490 | static int message_box(const char* title, const char* message); |
duke@435 | 491 | static char* do_you_want_to_debug(const char* message); |
duke@435 | 492 | |
duke@435 | 493 | // run cmd in a separate process and return its exit code; or -1 on failures |
duke@435 | 494 | static int fork_and_exec(char *cmd); |
duke@435 | 495 | |
duke@435 | 496 | // os::exit() is merged with vm_exit() |
duke@435 | 497 | // static void exit(int num); |
duke@435 | 498 | |
duke@435 | 499 | // Terminate the VM, but don't exit the process |
duke@435 | 500 | static void shutdown(); |
duke@435 | 501 | |
duke@435 | 502 | // Terminate with an error. Default is to generate a core file on platforms |
duke@435 | 503 | // that support such things. This calls shutdown() and then aborts. |
duke@435 | 504 | static void abort(bool dump_core = true); |
duke@435 | 505 | |
duke@435 | 506 | // Die immediately, no exit hook, no abort hook, no cleanup. |
duke@435 | 507 | static void die(); |
duke@435 | 508 | |
ikrylov@2322 | 509 | // File i/o operations |
ikrylov@2322 | 510 | static const int default_file_open_flags(); |
ikrylov@2322 | 511 | static int open(const char *path, int oflag, int mode); |
vlivanov@5027 | 512 | static FILE* open(int fd, const char* mode); |
ikrylov@2322 | 513 | static int close(int fd); |
ikrylov@2322 | 514 | static jlong lseek(int fd, jlong offset, int whence); |
ikrylov@2322 | 515 | static char* native_path(char *path); |
ikrylov@2322 | 516 | static int ftruncate(int fd, jlong length); |
ikrylov@2322 | 517 | static int fsync(int fd); |
ikrylov@2322 | 518 | static int available(int fd, jlong *bytes); |
ikrylov@2322 | 519 | |
ikrylov@2322 | 520 | //File i/o operations |
ikrylov@2322 | 521 | |
ikrylov@2322 | 522 | static size_t read(int fd, void *buf, unsigned int nBytes); |
ikrylov@2322 | 523 | static size_t restartable_read(int fd, void *buf, unsigned int nBytes); |
ikrylov@2322 | 524 | static size_t write(int fd, const void *buf, unsigned int nBytes); |
ikrylov@2322 | 525 | |
duke@435 | 526 | // Reading directories. |
duke@435 | 527 | static DIR* opendir(const char* dirname); |
duke@435 | 528 | static int readdir_buf_size(const char *path); |
duke@435 | 529 | static struct dirent* readdir(DIR* dirp, dirent* dbuf); |
duke@435 | 530 | static int closedir(DIR* dirp); |
duke@435 | 531 | |
duke@435 | 532 | // Dynamic library extension |
duke@435 | 533 | static const char* dll_file_extension(); |
duke@435 | 534 | |
duke@435 | 535 | static const char* get_temp_directory(); |
vlivanov@5027 | 536 | static const char* get_current_directory(char *buf, size_t buflen); |
duke@435 | 537 | |
kamg@677 | 538 | // Builds a platform-specific full library path given a ld path and lib name |
bpittore@4261 | 539 | // Returns true if buffer contains full path to existing file, false otherwise |
bpittore@4261 | 540 | static bool dll_build_name(char* buffer, size_t size, |
kamg@677 | 541 | const char* pathname, const char* fname); |
kamg@677 | 542 | |
duke@435 | 543 | // Symbol lookup, find nearest function name; basically it implements |
duke@435 | 544 | // dladdr() for all platforms. Name of the nearest function is copied |
dcubed@5365 | 545 | // to buf. Distance from its base address is optionally returned as offset. |
duke@435 | 546 | // If function name is not found, buf[0] is set to '\0' and offset is |
dcubed@5365 | 547 | // set to -1 (if offset is non-NULL). |
duke@435 | 548 | static bool dll_address_to_function_name(address addr, char* buf, |
duke@435 | 549 | int buflen, int* offset); |
duke@435 | 550 | |
duke@435 | 551 | // Locate DLL/DSO. On success, full path of the library is copied to |
dcubed@5365 | 552 | // buf, and offset is optionally set to be the distance between addr |
dcubed@5365 | 553 | // and the library's base address. On failure, buf[0] is set to '\0' |
dcubed@5365 | 554 | // and offset is set to -1 (if offset is non-NULL). |
duke@435 | 555 | static bool dll_address_to_library_name(address addr, char* buf, |
duke@435 | 556 | int buflen, int* offset); |
duke@435 | 557 | |
duke@435 | 558 | // Find out whether the pc is in the static code for jvm.dll/libjvm.so. |
duke@435 | 559 | static bool address_is_in_vm(address addr); |
duke@435 | 560 | |
duke@435 | 561 | // Loads .dll/.so and |
duke@435 | 562 | // in case of error it checks if .dll/.so was built for the |
duke@435 | 563 | // same architecture as Hotspot is running on |
duke@435 | 564 | static void* dll_load(const char *name, char *ebuf, int ebuflen); |
duke@435 | 565 | |
kamg@677 | 566 | // lookup symbol in a shared library |
kamg@677 | 567 | static void* dll_lookup(void* handle, const char* name); |
kamg@677 | 568 | |
ikrylov@2322 | 569 | // Unload library |
ikrylov@2322 | 570 | static void dll_unload(void *lib); |
ikrylov@2322 | 571 | |
bpittore@5585 | 572 | // Return the handle of this process |
bpittore@5585 | 573 | static void* get_default_process_handle(); |
bpittore@5585 | 574 | |
bpittore@5585 | 575 | // Check for static linked agent library |
bpittore@5585 | 576 | static bool find_builtin_agent(AgentLibrary *agent_lib, const char *syms[], |
bpittore@5585 | 577 | size_t syms_len); |
bpittore@5585 | 578 | |
bpittore@5585 | 579 | // Find agent entry point |
bpittore@5585 | 580 | static void *find_agent_function(AgentLibrary *agent_lib, bool check_lib, |
bpittore@5585 | 581 | const char *syms[], size_t syms_len); |
bpittore@5585 | 582 | |
duke@435 | 583 | // Print out system information; they are called by fatal error handler. |
duke@435 | 584 | // Output format may be different on different platforms. |
duke@435 | 585 | static void print_os_info(outputStream* st); |
nloodin@3783 | 586 | static void print_os_info_brief(outputStream* st); |
duke@435 | 587 | static void print_cpu_info(outputStream* st); |
jcoomes@2997 | 588 | static void pd_print_cpu_info(outputStream* st); |
duke@435 | 589 | static void print_memory_info(outputStream* st); |
duke@435 | 590 | static void print_dll_info(outputStream* st); |
duke@435 | 591 | static void print_environment_variables(outputStream* st, const char** env_list, char* buffer, int len); |
duke@435 | 592 | static void print_context(outputStream* st, void* context); |
never@2262 | 593 | static void print_register_info(outputStream* st, void* context); |
duke@435 | 594 | static void print_siginfo(outputStream* st, void* siginfo); |
duke@435 | 595 | static void print_signal_handlers(outputStream* st, char* buf, size_t buflen); |
duke@435 | 596 | static void print_date_and_time(outputStream* st); |
duke@435 | 597 | |
never@2262 | 598 | static void print_location(outputStream* st, intptr_t x, bool verbose = false); |
ikrylov@2322 | 599 | static size_t lasterror(char *buf, size_t len); |
phh@3379 | 600 | static int get_last_error(); |
bobv@2036 | 601 | |
sla@2584 | 602 | // Determines whether the calling process is being debugged by a user-mode debugger. |
sla@2584 | 603 | static bool is_debugger_attached(); |
sla@2584 | 604 | |
sla@2584 | 605 | // wait for a key press if PauseAtExit is set |
sla@2584 | 606 | static void wait_for_keypress_at_exit(void); |
sla@2584 | 607 | |
duke@435 | 608 | // The following two functions are used by fatal error handler to trace |
duke@435 | 609 | // native (C) frames. They are not part of frame.hpp/frame.cpp because |
duke@435 | 610 | // frame.hpp/cpp assume thread is JavaThread, and also because different |
duke@435 | 611 | // OS/compiler may have different convention or provide different API to |
duke@435 | 612 | // walk C frames. |
duke@435 | 613 | // |
duke@435 | 614 | // We don't attempt to become a debugger, so we only follow frames if that |
duke@435 | 615 | // does not require a lookup in the unwind table, which is part of the binary |
duke@435 | 616 | // file but may be unsafe to read after a fatal error. So on x86, we can |
duke@435 | 617 | // only walk stack if %ebp is used as frame pointer; on ia64, it's not |
duke@435 | 618 | // possible to walk C stack without having the unwind table. |
duke@435 | 619 | static bool is_first_C_frame(frame *fr); |
duke@435 | 620 | static frame get_sender_for_C_frame(frame *fr); |
duke@435 | 621 | |
duke@435 | 622 | // return current frame. pc() and sp() are set to NULL on failure. |
duke@435 | 623 | static frame current_frame(); |
duke@435 | 624 | |
duke@435 | 625 | static void print_hex_dump(outputStream* st, address start, address end, int unitsize); |
duke@435 | 626 | |
duke@435 | 627 | // returns a string to describe the exception/signal; |
duke@435 | 628 | // returns NULL if exception_code is not an OS exception/signal. |
duke@435 | 629 | static const char* exception_name(int exception_code, char* buf, size_t buflen); |
duke@435 | 630 | |
duke@435 | 631 | // Returns native Java library, loads if necessary |
duke@435 | 632 | static void* native_java_library(); |
duke@435 | 633 | |
ikrylov@2322 | 634 | // Fills in path to jvm.dll/libjvm.so (used by the Disassembler) |
duke@435 | 635 | static void jvm_path(char *buf, jint buflen); |
duke@435 | 636 | |
bobv@2036 | 637 | // Returns true if we are running in a headless jre. |
bobv@2036 | 638 | static bool is_headless_jre(); |
bobv@2036 | 639 | |
duke@435 | 640 | // JNI names |
duke@435 | 641 | static void print_jni_name_prefix_on(outputStream* st, int args_size); |
duke@435 | 642 | static void print_jni_name_suffix_on(outputStream* st, int args_size); |
duke@435 | 643 | |
duke@435 | 644 | // File conventions |
duke@435 | 645 | static const char* file_separator(); |
duke@435 | 646 | static const char* line_separator(); |
duke@435 | 647 | static const char* path_separator(); |
duke@435 | 648 | |
duke@435 | 649 | // Init os specific system properties values |
duke@435 | 650 | static void init_system_properties_values(); |
duke@435 | 651 | |
duke@435 | 652 | // IO operations, non-JVM_ version. |
duke@435 | 653 | static int stat(const char* path, struct stat* sbuf); |
duke@435 | 654 | static bool dir_is_empty(const char* path); |
duke@435 | 655 | |
duke@435 | 656 | // IO operations on binary files |
duke@435 | 657 | static int create_binary_file(const char* path, bool rewrite_existing); |
duke@435 | 658 | static jlong current_file_offset(int fd); |
duke@435 | 659 | static jlong seek_to_file_offset(int fd, jlong offset); |
duke@435 | 660 | |
duke@435 | 661 | // Thread Local Storage |
duke@435 | 662 | static int allocate_thread_local_storage(); |
duke@435 | 663 | static void thread_local_storage_at_put(int index, void* value); |
duke@435 | 664 | static void* thread_local_storage_at(int index); |
duke@435 | 665 | static void free_thread_local_storage(int index); |
duke@435 | 666 | |
zgu@7074 | 667 | // Retrieve native stack frames. |
zgu@7074 | 668 | // Parameter: |
zgu@7074 | 669 | // stack: an array to storage stack pointers. |
zgu@7074 | 670 | // frames: size of above array. |
zgu@7074 | 671 | // toSkip: number of stack frames to skip at the beginning. |
zgu@7074 | 672 | // Return: number of stack frames captured. |
zgu@7074 | 673 | static int get_native_stack(address* stack, int size, int toSkip = 0); |
zgu@3900 | 674 | |
duke@435 | 675 | // General allocation (must be MT-safe) |
zgu@7074 | 676 | static void* malloc (size_t size, MEMFLAGS flags, const NativeCallStack& stack); |
zgu@7074 | 677 | static void* malloc (size_t size, MEMFLAGS flags); |
zgu@7074 | 678 | static void* realloc (void *memblock, size_t size, MEMFLAGS flag, const NativeCallStack& stack); |
zgu@7074 | 679 | static void* realloc (void *memblock, size_t size, MEMFLAGS flag); |
zgu@7074 | 680 | |
zgu@3900 | 681 | static void free (void *memblock, MEMFLAGS flags = mtNone); |
duke@435 | 682 | static bool check_heap(bool force = false); // verify C heap integrity |
zgu@3900 | 683 | static char* strdup(const char *, MEMFLAGS flags = mtInternal); // Like strdup |
duke@435 | 684 | |
duke@435 | 685 | #ifndef PRODUCT |
kvn@2557 | 686 | static julong num_mallocs; // # of calls to malloc/realloc |
kvn@2557 | 687 | static julong alloc_bytes; // # of bytes allocated |
kvn@2557 | 688 | static julong num_frees; // # of calls to free |
kvn@2557 | 689 | static julong free_bytes; // # of bytes freed |
duke@435 | 690 | #endif |
duke@435 | 691 | |
ikrylov@2322 | 692 | // SocketInterface (ex HPI SocketInterface ) |
ikrylov@2322 | 693 | static int socket(int domain, int type, int protocol); |
ikrylov@2322 | 694 | static int socket_close(int fd); |
ikrylov@2322 | 695 | static int socket_shutdown(int fd, int howto); |
phh@3344 | 696 | static int recv(int fd, char* buf, size_t nBytes, uint flags); |
phh@3344 | 697 | static int send(int fd, char* buf, size_t nBytes, uint flags); |
phh@3344 | 698 | static int raw_send(int fd, char* buf, size_t nBytes, uint flags); |
ikrylov@2322 | 699 | static int timeout(int fd, long timeout); |
ikrylov@2322 | 700 | static int listen(int fd, int count); |
phh@3344 | 701 | static int connect(int fd, struct sockaddr* him, socklen_t len); |
phh@3344 | 702 | static int bind(int fd, struct sockaddr* him, socklen_t len); |
phh@3344 | 703 | static int accept(int fd, struct sockaddr* him, socklen_t* len); |
phh@3344 | 704 | static int recvfrom(int fd, char* buf, size_t nbytes, uint flags, |
phh@3344 | 705 | struct sockaddr* from, socklen_t* fromlen); |
phh@3344 | 706 | static int get_sock_name(int fd, struct sockaddr* him, socklen_t* len); |
phh@3344 | 707 | static int sendto(int fd, char* buf, size_t len, uint flags, |
phh@3344 | 708 | struct sockaddr* to, socklen_t tolen); |
phh@3344 | 709 | static int socket_available(int fd, jint* pbytes); |
ikrylov@2322 | 710 | |
ikrylov@2322 | 711 | static int get_sock_opt(int fd, int level, int optname, |
phh@3344 | 712 | char* optval, socklen_t* optlen); |
ikrylov@2322 | 713 | static int set_sock_opt(int fd, int level, int optname, |
phh@3344 | 714 | const char* optval, socklen_t optlen); |
ikrylov@2322 | 715 | static int get_host_name(char* name, int namelen); |
ikrylov@2322 | 716 | |
phh@3344 | 717 | static struct hostent* get_host_by_name(char* name); |
ikrylov@2322 | 718 | |
duke@435 | 719 | // Support for signals (see JVM_RaiseSignal, JVM_RegisterSignal) |
duke@435 | 720 | static void signal_init(); |
duke@435 | 721 | static void signal_init_pd(); |
duke@435 | 722 | static void signal_notify(int signal_number); |
duke@435 | 723 | static void* signal(int signal_number, void* handler); |
duke@435 | 724 | static void signal_raise(int signal_number); |
duke@435 | 725 | static int signal_wait(); |
duke@435 | 726 | static int signal_lookup(); |
duke@435 | 727 | static void* user_handler(); |
duke@435 | 728 | static void terminate_signal_thread(); |
duke@435 | 729 | static int sigexitnum_pd(); |
duke@435 | 730 | |
duke@435 | 731 | // random number generation |
duke@435 | 732 | static long random(); // return 32bit pseudorandom number |
duke@435 | 733 | static void init_random(long initval); // initialize random sequence |
duke@435 | 734 | |
duke@435 | 735 | // Structured OS Exception support |
duke@435 | 736 | static void os_exception_wrapper(java_call_t f, JavaValue* value, methodHandle* method, JavaCallArguments* args, Thread* thread); |
duke@435 | 737 | |
ctornqvi@2520 | 738 | // On Windows this will create an actual minidump, on Linux/Solaris it will simply check core dump limits |
ctornqvi@2520 | 739 | static void check_or_create_dump(void* exceptionRecord, void* contextRecord, char* buffer, size_t bufferSize); |
ctornqvi@2520 | 740 | |
mikael@3903 | 741 | // Get the default path to the core file |
mikael@3903 | 742 | // Returns the length of the string |
mikael@3903 | 743 | static int get_core_path(char* buffer, size_t bufferSize); |
mikael@3903 | 744 | |
duke@435 | 745 | // JVMTI & JVM monitoring and management support |
duke@435 | 746 | // The thread_cpu_time() and current_thread_cpu_time() are only |
duke@435 | 747 | // supported if is_thread_cpu_time_supported() returns true. |
duke@435 | 748 | // They are not supported on Solaris T1. |
duke@435 | 749 | |
duke@435 | 750 | // Thread CPU Time - return the fast estimate on a platform |
duke@435 | 751 | // On Solaris - call gethrvtime (fast) - user time only |
duke@435 | 752 | // On Linux - fast clock_gettime where available - user+sys |
duke@435 | 753 | // - otherwise: very slow /proc fs - user+sys |
duke@435 | 754 | // On Windows - GetThreadTimes - user+sys |
duke@435 | 755 | static jlong current_thread_cpu_time(); |
duke@435 | 756 | static jlong thread_cpu_time(Thread* t); |
duke@435 | 757 | |
duke@435 | 758 | // Thread CPU Time with user_sys_cpu_time parameter. |
duke@435 | 759 | // |
duke@435 | 760 | // If user_sys_cpu_time is true, user+sys time is returned. |
duke@435 | 761 | // Otherwise, only user time is returned |
duke@435 | 762 | static jlong current_thread_cpu_time(bool user_sys_cpu_time); |
duke@435 | 763 | static jlong thread_cpu_time(Thread* t, bool user_sys_cpu_time); |
duke@435 | 764 | |
duke@435 | 765 | // Return a bunch of info about the timers. |
duke@435 | 766 | // Note that the returned info for these two functions may be different |
duke@435 | 767 | // on some platforms |
duke@435 | 768 | static void current_thread_cpu_time_info(jvmtiTimerInfo *info_ptr); |
duke@435 | 769 | static void thread_cpu_time_info(jvmtiTimerInfo *info_ptr); |
duke@435 | 770 | |
duke@435 | 771 | static bool is_thread_cpu_time_supported(); |
duke@435 | 772 | |
duke@435 | 773 | // System loadavg support. Returns -1 if load average cannot be obtained. |
duke@435 | 774 | static int loadavg(double loadavg[], int nelem); |
duke@435 | 775 | |
duke@435 | 776 | // Hook for os specific jvm options that we don't want to abort on seeing |
duke@435 | 777 | static bool obsolete_option(const JavaVMOption *option); |
duke@435 | 778 | |
phh@3378 | 779 | // Extensions |
phh@3378 | 780 | #include "runtime/os_ext.hpp" |
phh@3378 | 781 | |
phh@3378 | 782 | public: |
rbackman@5424 | 783 | class CrashProtectionCallback : public StackObj { |
rbackman@5424 | 784 | public: |
rbackman@5424 | 785 | virtual void call() = 0; |
rbackman@5424 | 786 | }; |
phh@3378 | 787 | |
duke@435 | 788 | // Platform dependent stuff |
stefank@2314 | 789 | #ifdef TARGET_OS_FAMILY_linux |
stefank@2314 | 790 | # include "os_linux.hpp" |
nloodin@3783 | 791 | # include "os_posix.hpp" |
stefank@2314 | 792 | #endif |
stefank@2314 | 793 | #ifdef TARGET_OS_FAMILY_solaris |
stefank@2314 | 794 | # include "os_solaris.hpp" |
nloodin@3783 | 795 | # include "os_posix.hpp" |
stefank@2314 | 796 | #endif |
stefank@2314 | 797 | #ifdef TARGET_OS_FAMILY_windows |
stefank@2314 | 798 | # include "os_windows.hpp" |
stefank@2314 | 799 | #endif |
goetz@6461 | 800 | #ifdef TARGET_OS_FAMILY_aix |
goetz@6461 | 801 | # include "os_aix.hpp" |
goetz@6461 | 802 | # include "os_posix.hpp" |
goetz@6461 | 803 | #endif |
never@3156 | 804 | #ifdef TARGET_OS_FAMILY_bsd |
nloodin@3783 | 805 | # include "os_posix.hpp" |
never@3156 | 806 | # include "os_bsd.hpp" |
never@3156 | 807 | #endif |
stefank@2314 | 808 | #ifdef TARGET_OS_ARCH_linux_x86 |
stefank@2314 | 809 | # include "os_linux_x86.hpp" |
stefank@2314 | 810 | #endif |
stefank@2314 | 811 | #ifdef TARGET_OS_ARCH_linux_sparc |
stefank@2314 | 812 | # include "os_linux_sparc.hpp" |
stefank@2314 | 813 | #endif |
stefank@2314 | 814 | #ifdef TARGET_OS_ARCH_linux_zero |
stefank@2314 | 815 | # include "os_linux_zero.hpp" |
stefank@2314 | 816 | #endif |
stefank@2314 | 817 | #ifdef TARGET_OS_ARCH_solaris_x86 |
stefank@2314 | 818 | # include "os_solaris_x86.hpp" |
stefank@2314 | 819 | #endif |
stefank@2314 | 820 | #ifdef TARGET_OS_ARCH_solaris_sparc |
stefank@2314 | 821 | # include "os_solaris_sparc.hpp" |
stefank@2314 | 822 | #endif |
stefank@2314 | 823 | #ifdef TARGET_OS_ARCH_windows_x86 |
stefank@2314 | 824 | # include "os_windows_x86.hpp" |
stefank@2314 | 825 | #endif |
bobv@2508 | 826 | #ifdef TARGET_OS_ARCH_linux_arm |
bobv@2508 | 827 | # include "os_linux_arm.hpp" |
bobv@2508 | 828 | #endif |
bobv@2508 | 829 | #ifdef TARGET_OS_ARCH_linux_ppc |
bobv@2508 | 830 | # include "os_linux_ppc.hpp" |
bobv@2508 | 831 | #endif |
goetz@6461 | 832 | #ifdef TARGET_OS_ARCH_aix_ppc |
goetz@6461 | 833 | # include "os_aix_ppc.hpp" |
goetz@6461 | 834 | #endif |
never@3156 | 835 | #ifdef TARGET_OS_ARCH_bsd_x86 |
never@3156 | 836 | # include "os_bsd_x86.hpp" |
never@3156 | 837 | #endif |
never@3156 | 838 | #ifdef TARGET_OS_ARCH_bsd_zero |
never@3156 | 839 | # include "os_bsd_zero.hpp" |
never@3156 | 840 | #endif |
stefank@2314 | 841 | |
phh@3378 | 842 | public: |
iklam@5667 | 843 | #ifndef PLATFORM_PRINT_NATIVE_STACK |
iklam@5667 | 844 | // No platform-specific code for printing the native stack. |
iklam@5667 | 845 | static bool platform_print_native_stack(outputStream* st, void* context, |
iklam@5667 | 846 | char *buf, int buf_size) { |
iklam@5667 | 847 | return false; |
iklam@5667 | 848 | } |
iklam@5667 | 849 | #endif |
iklam@5667 | 850 | |
bobv@2036 | 851 | // debugging support (mostly used by debug.cpp but also fatal error handler) |
bobv@2036 | 852 | static bool find(address pc, outputStream* st = tty); // OS specific function to make sense out of an address |
duke@435 | 853 | |
duke@435 | 854 | static bool dont_yield(); // when true, JVM_Yield() is nop |
duke@435 | 855 | static void print_statistics(); |
duke@435 | 856 | |
duke@435 | 857 | // Thread priority helpers (implemented in OS-specific part) |
duke@435 | 858 | static OSReturn set_native_priority(Thread* thread, int native_prio); |
duke@435 | 859 | static OSReturn get_native_priority(const Thread* const thread, int* priority_ptr); |
phh@3481 | 860 | static int java_to_os_priority[CriticalPriority + 1]; |
duke@435 | 861 | // Hint to the underlying OS that a task switch would not be good. |
duke@435 | 862 | // Void return because it's a hint and can fail. |
duke@435 | 863 | static void hint_no_preempt(); |
duke@435 | 864 | |
duke@435 | 865 | // Used at creation if requested by the diagnostic flag PauseAtStartup. |
duke@435 | 866 | // Causes the VM to wait until an external stimulus has been applied |
duke@435 | 867 | // (for Unix, that stimulus is a signal, for Windows, an external |
duke@435 | 868 | // ResumeThread call) |
duke@435 | 869 | static void pause(); |
duke@435 | 870 | |
bpittore@5585 | 871 | // Builds a platform dependent Agent_OnLoad_<libname> function name |
bpittore@5585 | 872 | // which is used to find statically linked in agents. |
bpittore@5585 | 873 | static char* build_agent_function_name(const char *sym, const char *cname, |
bpittore@5585 | 874 | bool is_absolute_path); |
bpittore@5585 | 875 | |
sla@5237 | 876 | class SuspendedThreadTaskContext { |
sla@5237 | 877 | public: |
sla@5237 | 878 | SuspendedThreadTaskContext(Thread* thread, void *ucontext) : _thread(thread), _ucontext(ucontext) {} |
sla@5237 | 879 | Thread* thread() const { return _thread; } |
sla@5237 | 880 | void* ucontext() const { return _ucontext; } |
sla@5237 | 881 | private: |
sla@5237 | 882 | Thread* _thread; |
sla@5237 | 883 | void* _ucontext; |
sla@5237 | 884 | }; |
sla@5237 | 885 | |
sla@5237 | 886 | class SuspendedThreadTask { |
sla@5237 | 887 | public: |
sla@5237 | 888 | SuspendedThreadTask(Thread* thread) : _thread(thread), _done(false) {} |
sla@5237 | 889 | virtual ~SuspendedThreadTask() {} |
sla@5237 | 890 | void run(); |
sla@5237 | 891 | bool is_done() { return _done; } |
sla@5237 | 892 | virtual void do_task(const SuspendedThreadTaskContext& context) = 0; |
sla@5237 | 893 | protected: |
sla@5237 | 894 | private: |
sla@5237 | 895 | void internal_do_task(); |
sla@5237 | 896 | Thread* _thread; |
sla@5237 | 897 | bool _done; |
sla@5237 | 898 | }; |
sla@5237 | 899 | |
sla@5237 | 900 | #ifndef TARGET_OS_FAMILY_windows |
sla@5237 | 901 | // Suspend/resume support |
sla@5237 | 902 | // Protocol: |
sla@5237 | 903 | // |
sla@5237 | 904 | // a thread starts in SR_RUNNING |
sla@5237 | 905 | // |
sla@5237 | 906 | // SR_RUNNING can go to |
sla@5237 | 907 | // * SR_SUSPEND_REQUEST when the WatcherThread wants to suspend it |
sla@5237 | 908 | // SR_SUSPEND_REQUEST can go to |
sla@5237 | 909 | // * SR_RUNNING if WatcherThread decides it waited for SR_SUSPENDED too long (timeout) |
sla@5237 | 910 | // * SR_SUSPENDED if the stopped thread receives the signal and switches state |
sla@5237 | 911 | // SR_SUSPENDED can go to |
sla@5237 | 912 | // * SR_WAKEUP_REQUEST when the WatcherThread has done the work and wants to resume |
sla@5237 | 913 | // SR_WAKEUP_REQUEST can go to |
sla@5237 | 914 | // * SR_RUNNING when the stopped thread receives the signal |
sla@5237 | 915 | // * SR_WAKEUP_REQUEST on timeout (resend the signal and try again) |
sla@5237 | 916 | class SuspendResume { |
sla@5237 | 917 | public: |
sla@5237 | 918 | enum State { |
sla@5237 | 919 | SR_RUNNING, |
sla@5237 | 920 | SR_SUSPEND_REQUEST, |
sla@5237 | 921 | SR_SUSPENDED, |
sla@5237 | 922 | SR_WAKEUP_REQUEST |
sla@5237 | 923 | }; |
sla@5237 | 924 | |
sla@5237 | 925 | private: |
sla@5237 | 926 | volatile State _state; |
sla@5237 | 927 | |
sla@5237 | 928 | private: |
sla@5237 | 929 | /* try to switch state from state "from" to state "to" |
sla@5237 | 930 | * returns the state set after the method is complete |
sla@5237 | 931 | */ |
sla@5237 | 932 | State switch_state(State from, State to); |
sla@5237 | 933 | |
sla@5237 | 934 | public: |
sla@5237 | 935 | SuspendResume() : _state(SR_RUNNING) { } |
sla@5237 | 936 | |
sla@5237 | 937 | State state() const { return _state; } |
sla@5237 | 938 | |
sla@5237 | 939 | State request_suspend() { |
sla@5237 | 940 | return switch_state(SR_RUNNING, SR_SUSPEND_REQUEST); |
sla@5237 | 941 | } |
sla@5237 | 942 | |
sla@5237 | 943 | State cancel_suspend() { |
sla@5237 | 944 | return switch_state(SR_SUSPEND_REQUEST, SR_RUNNING); |
sla@5237 | 945 | } |
sla@5237 | 946 | |
sla@5237 | 947 | State suspended() { |
sla@5237 | 948 | return switch_state(SR_SUSPEND_REQUEST, SR_SUSPENDED); |
sla@5237 | 949 | } |
sla@5237 | 950 | |
sla@5237 | 951 | State request_wakeup() { |
sla@5237 | 952 | return switch_state(SR_SUSPENDED, SR_WAKEUP_REQUEST); |
sla@5237 | 953 | } |
sla@5237 | 954 | |
sla@5237 | 955 | State running() { |
sla@5237 | 956 | return switch_state(SR_WAKEUP_REQUEST, SR_RUNNING); |
sla@5237 | 957 | } |
sla@5237 | 958 | |
sla@5237 | 959 | bool is_running() const { |
sla@5237 | 960 | return _state == SR_RUNNING; |
sla@5237 | 961 | } |
sla@5237 | 962 | |
sla@5237 | 963 | bool is_suspend_request() const { |
sla@5237 | 964 | return _state == SR_SUSPEND_REQUEST; |
sla@5237 | 965 | } |
sla@5237 | 966 | |
sla@5237 | 967 | bool is_suspended() const { |
sla@5237 | 968 | return _state == SR_SUSPENDED; |
sla@5237 | 969 | } |
sla@5237 | 970 | }; |
sla@5237 | 971 | #endif |
sla@5237 | 972 | |
sla@5237 | 973 | |
duke@435 | 974 | protected: |
duke@435 | 975 | static long _rand_seed; // seed for random number generator |
duke@435 | 976 | static int _processor_count; // number of processors |
duke@435 | 977 | |
duke@435 | 978 | static char* format_boot_path(const char* format_string, |
duke@435 | 979 | const char* home, |
duke@435 | 980 | int home_len, |
duke@435 | 981 | char fileSep, |
duke@435 | 982 | char pathSep); |
duke@435 | 983 | static bool set_boot_path(char fileSep, char pathSep); |
phh@1126 | 984 | static char** split_path(const char* path, int* n); |
rbackman@5424 | 985 | |
duke@435 | 986 | }; |
duke@435 | 987 | |
duke@435 | 988 | // Note that "PAUSE" is almost always used with synchronization |
duke@435 | 989 | // so arguably we should provide Atomic::SpinPause() instead |
duke@435 | 990 | // of the global SpinPause() with C linkage. |
duke@435 | 991 | // It'd also be eligible for inlining on many platforms. |
duke@435 | 992 | |
goetz@5400 | 993 | extern "C" int SpinPause(); |
stefank@2314 | 994 | |
stefank@2314 | 995 | #endif // SHARE_VM_RUNTIME_OS_HPP |