Mon, 30 Jun 2008 17:04:59 -0700
6618726: Introduce -XX:+UnlockExperimentalVMOptions flag
Summary: experimental() flags will protect features of an experimental nature that are not supported in the regular product build. Made UseG1GC an experimental flag.
Reviewed-by: jmasa, kamg, coleenp
duke@435 | 1 | /* |
duke@435 | 2 | * Copyright 1997-2007 Sun Microsystems, Inc. All Rights Reserved. |
duke@435 | 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
duke@435 | 4 | * |
duke@435 | 5 | * This code is free software; you can redistribute it and/or modify it |
duke@435 | 6 | * under the terms of the GNU General Public License version 2 only, as |
duke@435 | 7 | * published by the Free Software Foundation. |
duke@435 | 8 | * |
duke@435 | 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
duke@435 | 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
duke@435 | 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
duke@435 | 12 | * version 2 for more details (a copy is included in the LICENSE file that |
duke@435 | 13 | * accompanied this code). |
duke@435 | 14 | * |
duke@435 | 15 | * You should have received a copy of the GNU General Public License version |
duke@435 | 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
duke@435 | 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
duke@435 | 18 | * |
duke@435 | 19 | * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara, |
duke@435 | 20 | * CA 95054 USA or visit www.sun.com if you need additional information or |
duke@435 | 21 | * have any questions. |
duke@435 | 22 | * |
duke@435 | 23 | */ |
duke@435 | 24 | |
duke@435 | 25 | # include "incls/_precompiled.incl" |
duke@435 | 26 | # include "incls/_os.cpp.incl" |
duke@435 | 27 | |
duke@435 | 28 | # include <signal.h> |
duke@435 | 29 | |
duke@435 | 30 | OSThread* os::_starting_thread = NULL; |
duke@435 | 31 | address os::_polling_page = NULL; |
duke@435 | 32 | volatile int32_t* os::_mem_serialize_page = NULL; |
duke@435 | 33 | uintptr_t os::_serialize_page_mask = 0; |
duke@435 | 34 | long os::_rand_seed = 1; |
duke@435 | 35 | int os::_processor_count = 0; |
duke@435 | 36 | size_t os::_page_sizes[os::page_sizes_max]; |
duke@435 | 37 | |
duke@435 | 38 | #ifndef PRODUCT |
duke@435 | 39 | int os::num_mallocs = 0; // # of calls to malloc/realloc |
duke@435 | 40 | size_t os::alloc_bytes = 0; // # of bytes allocated |
duke@435 | 41 | int os::num_frees = 0; // # of calls to free |
duke@435 | 42 | #endif |
duke@435 | 43 | |
duke@435 | 44 | // Fill in buffer with current local time as an ISO-8601 string. |
duke@435 | 45 | // E.g., yyyy-mm-ddThh:mm:ss-zzzz. |
duke@435 | 46 | // Returns buffer, or NULL if it failed. |
duke@435 | 47 | // This would mostly be a call to |
duke@435 | 48 | // strftime(...., "%Y-%m-%d" "T" "%H:%M:%S" "%z", ....) |
duke@435 | 49 | // except that on Windows the %z behaves badly, so we do it ourselves. |
duke@435 | 50 | // Also, people wanted milliseconds on there, |
duke@435 | 51 | // and strftime doesn't do milliseconds. |
duke@435 | 52 | char* os::iso8601_time(char* buffer, size_t buffer_length) { |
duke@435 | 53 | // Output will be of the form "YYYY-MM-DDThh:mm:ss.mmm+zzzz\0" |
duke@435 | 54 | // 1 2 |
duke@435 | 55 | // 12345678901234567890123456789 |
duke@435 | 56 | static const char* iso8601_format = |
duke@435 | 57 | "%04d-%02d-%02dT%02d:%02d:%02d.%03d%c%02d%02d"; |
duke@435 | 58 | static const size_t needed_buffer = 29; |
duke@435 | 59 | |
duke@435 | 60 | // Sanity check the arguments |
duke@435 | 61 | if (buffer == NULL) { |
duke@435 | 62 | assert(false, "NULL buffer"); |
duke@435 | 63 | return NULL; |
duke@435 | 64 | } |
duke@435 | 65 | if (buffer_length < needed_buffer) { |
duke@435 | 66 | assert(false, "buffer_length too small"); |
duke@435 | 67 | return NULL; |
duke@435 | 68 | } |
duke@435 | 69 | // Get the current time |
sbohne@496 | 70 | jlong milliseconds_since_19700101 = javaTimeMillis(); |
duke@435 | 71 | const int milliseconds_per_microsecond = 1000; |
duke@435 | 72 | const time_t seconds_since_19700101 = |
duke@435 | 73 | milliseconds_since_19700101 / milliseconds_per_microsecond; |
duke@435 | 74 | const int milliseconds_after_second = |
duke@435 | 75 | milliseconds_since_19700101 % milliseconds_per_microsecond; |
duke@435 | 76 | // Convert the time value to a tm and timezone variable |
duke@435 | 77 | const struct tm *time_struct_temp = localtime(&seconds_since_19700101); |
duke@435 | 78 | if (time_struct_temp == NULL) { |
duke@435 | 79 | assert(false, "Failed localtime"); |
duke@435 | 80 | return NULL; |
duke@435 | 81 | } |
duke@435 | 82 | // Save the results of localtime |
duke@435 | 83 | const struct tm time_struct = *time_struct_temp; |
duke@435 | 84 | const time_t zone = timezone; |
duke@435 | 85 | |
duke@435 | 86 | // If daylight savings time is in effect, |
duke@435 | 87 | // we are 1 hour East of our time zone |
duke@435 | 88 | const time_t seconds_per_minute = 60; |
duke@435 | 89 | const time_t minutes_per_hour = 60; |
duke@435 | 90 | const time_t seconds_per_hour = seconds_per_minute * minutes_per_hour; |
duke@435 | 91 | time_t UTC_to_local = zone; |
duke@435 | 92 | if (time_struct.tm_isdst > 0) { |
duke@435 | 93 | UTC_to_local = UTC_to_local - seconds_per_hour; |
duke@435 | 94 | } |
duke@435 | 95 | // Compute the time zone offset. |
duke@435 | 96 | // localtime(3C) sets timezone to the difference (in seconds) |
duke@435 | 97 | // between UTC and and local time. |
duke@435 | 98 | // ISO 8601 says we need the difference between local time and UTC, |
duke@435 | 99 | // we change the sign of the localtime(3C) result. |
duke@435 | 100 | const time_t local_to_UTC = -(UTC_to_local); |
duke@435 | 101 | // Then we have to figure out if if we are ahead (+) or behind (-) UTC. |
duke@435 | 102 | char sign_local_to_UTC = '+'; |
duke@435 | 103 | time_t abs_local_to_UTC = local_to_UTC; |
duke@435 | 104 | if (local_to_UTC < 0) { |
duke@435 | 105 | sign_local_to_UTC = '-'; |
duke@435 | 106 | abs_local_to_UTC = -(abs_local_to_UTC); |
duke@435 | 107 | } |
duke@435 | 108 | // Convert time zone offset seconds to hours and minutes. |
duke@435 | 109 | const time_t zone_hours = (abs_local_to_UTC / seconds_per_hour); |
duke@435 | 110 | const time_t zone_min = |
duke@435 | 111 | ((abs_local_to_UTC % seconds_per_hour) / seconds_per_minute); |
duke@435 | 112 | |
duke@435 | 113 | // Print an ISO 8601 date and time stamp into the buffer |
duke@435 | 114 | const int year = 1900 + time_struct.tm_year; |
duke@435 | 115 | const int month = 1 + time_struct.tm_mon; |
duke@435 | 116 | const int printed = jio_snprintf(buffer, buffer_length, iso8601_format, |
duke@435 | 117 | year, |
duke@435 | 118 | month, |
duke@435 | 119 | time_struct.tm_mday, |
duke@435 | 120 | time_struct.tm_hour, |
duke@435 | 121 | time_struct.tm_min, |
duke@435 | 122 | time_struct.tm_sec, |
duke@435 | 123 | milliseconds_after_second, |
duke@435 | 124 | sign_local_to_UTC, |
duke@435 | 125 | zone_hours, |
duke@435 | 126 | zone_min); |
duke@435 | 127 | if (printed == 0) { |
duke@435 | 128 | assert(false, "Failed jio_printf"); |
duke@435 | 129 | return NULL; |
duke@435 | 130 | } |
duke@435 | 131 | return buffer; |
duke@435 | 132 | } |
duke@435 | 133 | |
duke@435 | 134 | OSReturn os::set_priority(Thread* thread, ThreadPriority p) { |
duke@435 | 135 | #ifdef ASSERT |
duke@435 | 136 | if (!(!thread->is_Java_thread() || |
duke@435 | 137 | Thread::current() == thread || |
duke@435 | 138 | Threads_lock->owned_by_self() |
duke@435 | 139 | || thread->is_Compiler_thread() |
duke@435 | 140 | )) { |
duke@435 | 141 | assert(false, "possibility of dangling Thread pointer"); |
duke@435 | 142 | } |
duke@435 | 143 | #endif |
duke@435 | 144 | |
duke@435 | 145 | if (p >= MinPriority && p <= MaxPriority) { |
duke@435 | 146 | int priority = java_to_os_priority[p]; |
duke@435 | 147 | return set_native_priority(thread, priority); |
duke@435 | 148 | } else { |
duke@435 | 149 | assert(false, "Should not happen"); |
duke@435 | 150 | return OS_ERR; |
duke@435 | 151 | } |
duke@435 | 152 | } |
duke@435 | 153 | |
duke@435 | 154 | |
duke@435 | 155 | OSReturn os::get_priority(const Thread* const thread, ThreadPriority& priority) { |
duke@435 | 156 | int p; |
duke@435 | 157 | int os_prio; |
duke@435 | 158 | OSReturn ret = get_native_priority(thread, &os_prio); |
duke@435 | 159 | if (ret != OS_OK) return ret; |
duke@435 | 160 | |
duke@435 | 161 | for (p = MaxPriority; p > MinPriority && java_to_os_priority[p] > os_prio; p--) ; |
duke@435 | 162 | priority = (ThreadPriority)p; |
duke@435 | 163 | return OS_OK; |
duke@435 | 164 | } |
duke@435 | 165 | |
duke@435 | 166 | |
duke@435 | 167 | // --------------------- sun.misc.Signal (optional) --------------------- |
duke@435 | 168 | |
duke@435 | 169 | |
duke@435 | 170 | // SIGBREAK is sent by the keyboard to query the VM state |
duke@435 | 171 | #ifndef SIGBREAK |
duke@435 | 172 | #define SIGBREAK SIGQUIT |
duke@435 | 173 | #endif |
duke@435 | 174 | |
duke@435 | 175 | // sigexitnum_pd is a platform-specific special signal used for terminating the Signal thread. |
duke@435 | 176 | |
duke@435 | 177 | |
duke@435 | 178 | static void signal_thread_entry(JavaThread* thread, TRAPS) { |
duke@435 | 179 | os::set_priority(thread, NearMaxPriority); |
duke@435 | 180 | while (true) { |
duke@435 | 181 | int sig; |
duke@435 | 182 | { |
duke@435 | 183 | // FIXME : Currently we have not decieded what should be the status |
duke@435 | 184 | // for this java thread blocked here. Once we decide about |
duke@435 | 185 | // that we should fix this. |
duke@435 | 186 | sig = os::signal_wait(); |
duke@435 | 187 | } |
duke@435 | 188 | if (sig == os::sigexitnum_pd()) { |
duke@435 | 189 | // Terminate the signal thread |
duke@435 | 190 | return; |
duke@435 | 191 | } |
duke@435 | 192 | |
duke@435 | 193 | switch (sig) { |
duke@435 | 194 | case SIGBREAK: { |
duke@435 | 195 | // Check if the signal is a trigger to start the Attach Listener - in that |
duke@435 | 196 | // case don't print stack traces. |
duke@435 | 197 | if (!DisableAttachMechanism && AttachListener::is_init_trigger()) { |
duke@435 | 198 | continue; |
duke@435 | 199 | } |
duke@435 | 200 | // Print stack traces |
duke@435 | 201 | // Any SIGBREAK operations added here should make sure to flush |
duke@435 | 202 | // the output stream (e.g. tty->flush()) after output. See 4803766. |
duke@435 | 203 | // Each module also prints an extra carriage return after its output. |
duke@435 | 204 | VM_PrintThreads op; |
duke@435 | 205 | VMThread::execute(&op); |
duke@435 | 206 | VM_PrintJNI jni_op; |
duke@435 | 207 | VMThread::execute(&jni_op); |
duke@435 | 208 | VM_FindDeadlocks op1(tty); |
duke@435 | 209 | VMThread::execute(&op1); |
duke@435 | 210 | Universe::print_heap_at_SIGBREAK(); |
duke@435 | 211 | if (PrintClassHistogram) { |
duke@435 | 212 | VM_GC_HeapInspection op1(gclog_or_tty, true /* force full GC before heap inspection */); |
duke@435 | 213 | VMThread::execute(&op1); |
duke@435 | 214 | } |
duke@435 | 215 | if (JvmtiExport::should_post_data_dump()) { |
duke@435 | 216 | JvmtiExport::post_data_dump(); |
duke@435 | 217 | } |
duke@435 | 218 | break; |
duke@435 | 219 | } |
duke@435 | 220 | default: { |
duke@435 | 221 | // Dispatch the signal to java |
duke@435 | 222 | HandleMark hm(THREAD); |
duke@435 | 223 | klassOop k = SystemDictionary::resolve_or_null(vmSymbolHandles::sun_misc_Signal(), THREAD); |
duke@435 | 224 | KlassHandle klass (THREAD, k); |
duke@435 | 225 | if (klass.not_null()) { |
duke@435 | 226 | JavaValue result(T_VOID); |
duke@435 | 227 | JavaCallArguments args; |
duke@435 | 228 | args.push_int(sig); |
duke@435 | 229 | JavaCalls::call_static( |
duke@435 | 230 | &result, |
duke@435 | 231 | klass, |
duke@435 | 232 | vmSymbolHandles::dispatch_name(), |
duke@435 | 233 | vmSymbolHandles::int_void_signature(), |
duke@435 | 234 | &args, |
duke@435 | 235 | THREAD |
duke@435 | 236 | ); |
duke@435 | 237 | } |
duke@435 | 238 | if (HAS_PENDING_EXCEPTION) { |
duke@435 | 239 | // tty is initialized early so we don't expect it to be null, but |
duke@435 | 240 | // if it is we can't risk doing an initialization that might |
duke@435 | 241 | // trigger additional out-of-memory conditions |
duke@435 | 242 | if (tty != NULL) { |
duke@435 | 243 | char klass_name[256]; |
duke@435 | 244 | char tmp_sig_name[16]; |
duke@435 | 245 | const char* sig_name = "UNKNOWN"; |
duke@435 | 246 | instanceKlass::cast(PENDING_EXCEPTION->klass())-> |
duke@435 | 247 | name()->as_klass_external_name(klass_name, 256); |
duke@435 | 248 | if (os::exception_name(sig, tmp_sig_name, 16) != NULL) |
duke@435 | 249 | sig_name = tmp_sig_name; |
duke@435 | 250 | warning("Exception %s occurred dispatching signal %s to handler" |
duke@435 | 251 | "- the VM may need to be forcibly terminated", |
duke@435 | 252 | klass_name, sig_name ); |
duke@435 | 253 | } |
duke@435 | 254 | CLEAR_PENDING_EXCEPTION; |
duke@435 | 255 | } |
duke@435 | 256 | } |
duke@435 | 257 | } |
duke@435 | 258 | } |
duke@435 | 259 | } |
duke@435 | 260 | |
duke@435 | 261 | |
duke@435 | 262 | void os::signal_init() { |
duke@435 | 263 | if (!ReduceSignalUsage) { |
duke@435 | 264 | // Setup JavaThread for processing signals |
duke@435 | 265 | EXCEPTION_MARK; |
duke@435 | 266 | klassOop k = SystemDictionary::resolve_or_fail(vmSymbolHandles::java_lang_Thread(), true, CHECK); |
duke@435 | 267 | instanceKlassHandle klass (THREAD, k); |
duke@435 | 268 | instanceHandle thread_oop = klass->allocate_instance_handle(CHECK); |
duke@435 | 269 | |
duke@435 | 270 | const char thread_name[] = "Signal Dispatcher"; |
duke@435 | 271 | Handle string = java_lang_String::create_from_str(thread_name, CHECK); |
duke@435 | 272 | |
duke@435 | 273 | // Initialize thread_oop to put it into the system threadGroup |
duke@435 | 274 | Handle thread_group (THREAD, Universe::system_thread_group()); |
duke@435 | 275 | JavaValue result(T_VOID); |
duke@435 | 276 | JavaCalls::call_special(&result, thread_oop, |
duke@435 | 277 | klass, |
duke@435 | 278 | vmSymbolHandles::object_initializer_name(), |
duke@435 | 279 | vmSymbolHandles::threadgroup_string_void_signature(), |
duke@435 | 280 | thread_group, |
duke@435 | 281 | string, |
duke@435 | 282 | CHECK); |
duke@435 | 283 | |
duke@435 | 284 | KlassHandle group(THREAD, SystemDictionary::threadGroup_klass()); |
duke@435 | 285 | JavaCalls::call_special(&result, |
duke@435 | 286 | thread_group, |
duke@435 | 287 | group, |
duke@435 | 288 | vmSymbolHandles::add_method_name(), |
duke@435 | 289 | vmSymbolHandles::thread_void_signature(), |
duke@435 | 290 | thread_oop, // ARG 1 |
duke@435 | 291 | CHECK); |
duke@435 | 292 | |
duke@435 | 293 | os::signal_init_pd(); |
duke@435 | 294 | |
duke@435 | 295 | { MutexLocker mu(Threads_lock); |
duke@435 | 296 | JavaThread* signal_thread = new JavaThread(&signal_thread_entry); |
duke@435 | 297 | |
duke@435 | 298 | // At this point it may be possible that no osthread was created for the |
duke@435 | 299 | // JavaThread due to lack of memory. We would have to throw an exception |
duke@435 | 300 | // in that case. However, since this must work and we do not allow |
duke@435 | 301 | // exceptions anyway, check and abort if this fails. |
duke@435 | 302 | if (signal_thread == NULL || signal_thread->osthread() == NULL) { |
duke@435 | 303 | vm_exit_during_initialization("java.lang.OutOfMemoryError", |
duke@435 | 304 | "unable to create new native thread"); |
duke@435 | 305 | } |
duke@435 | 306 | |
duke@435 | 307 | java_lang_Thread::set_thread(thread_oop(), signal_thread); |
duke@435 | 308 | java_lang_Thread::set_priority(thread_oop(), NearMaxPriority); |
duke@435 | 309 | java_lang_Thread::set_daemon(thread_oop()); |
duke@435 | 310 | |
duke@435 | 311 | signal_thread->set_threadObj(thread_oop()); |
duke@435 | 312 | Threads::add(signal_thread); |
duke@435 | 313 | Thread::start(signal_thread); |
duke@435 | 314 | } |
duke@435 | 315 | // Handle ^BREAK |
duke@435 | 316 | os::signal(SIGBREAK, os::user_handler()); |
duke@435 | 317 | } |
duke@435 | 318 | } |
duke@435 | 319 | |
duke@435 | 320 | |
duke@435 | 321 | void os::terminate_signal_thread() { |
duke@435 | 322 | if (!ReduceSignalUsage) |
duke@435 | 323 | signal_notify(sigexitnum_pd()); |
duke@435 | 324 | } |
duke@435 | 325 | |
duke@435 | 326 | |
duke@435 | 327 | // --------------------- loading libraries --------------------- |
duke@435 | 328 | |
duke@435 | 329 | typedef jint (JNICALL *JNI_OnLoad_t)(JavaVM *, void *); |
duke@435 | 330 | extern struct JavaVM_ main_vm; |
duke@435 | 331 | |
duke@435 | 332 | static void* _native_java_library = NULL; |
duke@435 | 333 | |
duke@435 | 334 | void* os::native_java_library() { |
duke@435 | 335 | if (_native_java_library == NULL) { |
duke@435 | 336 | char buffer[JVM_MAXPATHLEN]; |
duke@435 | 337 | char ebuf[1024]; |
duke@435 | 338 | |
duke@435 | 339 | // Try to load verify dll first. In 1.3 java dll depends on it and is not always |
duke@435 | 340 | // able to find it when the loading executable is outside the JDK. |
duke@435 | 341 | // In order to keep working with 1.2 we ignore any loading errors. |
duke@435 | 342 | hpi::dll_build_name(buffer, sizeof(buffer), Arguments::get_dll_dir(), "verify"); |
duke@435 | 343 | hpi::dll_load(buffer, ebuf, sizeof(ebuf)); |
duke@435 | 344 | |
duke@435 | 345 | // Load java dll |
duke@435 | 346 | hpi::dll_build_name(buffer, sizeof(buffer), Arguments::get_dll_dir(), "java"); |
duke@435 | 347 | _native_java_library = hpi::dll_load(buffer, ebuf, sizeof(ebuf)); |
duke@435 | 348 | if (_native_java_library == NULL) { |
duke@435 | 349 | vm_exit_during_initialization("Unable to load native library", ebuf); |
duke@435 | 350 | } |
duke@435 | 351 | // The JNI_OnLoad handling is normally done by method load in java.lang.ClassLoader$NativeLibrary, |
duke@435 | 352 | // but the VM loads the base library explicitly so we have to check for JNI_OnLoad as well |
duke@435 | 353 | const char *onLoadSymbols[] = JNI_ONLOAD_SYMBOLS; |
duke@435 | 354 | JNI_OnLoad_t JNI_OnLoad = CAST_TO_FN_PTR(JNI_OnLoad_t, hpi::dll_lookup(_native_java_library, onLoadSymbols[0])); |
duke@435 | 355 | if (JNI_OnLoad != NULL) { |
duke@435 | 356 | JavaThread* thread = JavaThread::current(); |
duke@435 | 357 | ThreadToNativeFromVM ttn(thread); |
duke@435 | 358 | HandleMark hm(thread); |
duke@435 | 359 | jint ver = (*JNI_OnLoad)(&main_vm, NULL); |
duke@435 | 360 | if (!Threads::is_supported_jni_version_including_1_1(ver)) { |
duke@435 | 361 | vm_exit_during_initialization("Unsupported JNI version"); |
duke@435 | 362 | } |
duke@435 | 363 | } |
duke@435 | 364 | } |
duke@435 | 365 | return _native_java_library; |
duke@435 | 366 | } |
duke@435 | 367 | |
duke@435 | 368 | // --------------------- heap allocation utilities --------------------- |
duke@435 | 369 | |
duke@435 | 370 | char *os::strdup(const char *str) { |
duke@435 | 371 | size_t size = strlen(str); |
duke@435 | 372 | char *dup_str = (char *)malloc(size + 1); |
duke@435 | 373 | if (dup_str == NULL) return NULL; |
duke@435 | 374 | strcpy(dup_str, str); |
duke@435 | 375 | return dup_str; |
duke@435 | 376 | } |
duke@435 | 377 | |
duke@435 | 378 | |
duke@435 | 379 | |
duke@435 | 380 | #ifdef ASSERT |
duke@435 | 381 | #define space_before (MallocCushion + sizeof(double)) |
duke@435 | 382 | #define space_after MallocCushion |
duke@435 | 383 | #define size_addr_from_base(p) (size_t*)(p + space_before - sizeof(size_t)) |
duke@435 | 384 | #define size_addr_from_obj(p) ((size_t*)p - 1) |
duke@435 | 385 | // MallocCushion: size of extra cushion allocated around objects with +UseMallocOnly |
duke@435 | 386 | // NB: cannot be debug variable, because these aren't set from the command line until |
duke@435 | 387 | // *after* the first few allocs already happened |
duke@435 | 388 | #define MallocCushion 16 |
duke@435 | 389 | #else |
duke@435 | 390 | #define space_before 0 |
duke@435 | 391 | #define space_after 0 |
duke@435 | 392 | #define size_addr_from_base(p) should not use w/o ASSERT |
duke@435 | 393 | #define size_addr_from_obj(p) should not use w/o ASSERT |
duke@435 | 394 | #define MallocCushion 0 |
duke@435 | 395 | #endif |
duke@435 | 396 | #define paranoid 0 /* only set to 1 if you suspect checking code has bug */ |
duke@435 | 397 | |
duke@435 | 398 | #ifdef ASSERT |
duke@435 | 399 | inline size_t get_size(void* obj) { |
duke@435 | 400 | size_t size = *size_addr_from_obj(obj); |
duke@435 | 401 | if (size < 0 ) |
duke@435 | 402 | fatal2("free: size field of object #%p was overwritten (%lu)", obj, size); |
duke@435 | 403 | return size; |
duke@435 | 404 | } |
duke@435 | 405 | |
duke@435 | 406 | u_char* find_cushion_backwards(u_char* start) { |
duke@435 | 407 | u_char* p = start; |
duke@435 | 408 | while (p[ 0] != badResourceValue || p[-1] != badResourceValue || |
duke@435 | 409 | p[-2] != badResourceValue || p[-3] != badResourceValue) p--; |
duke@435 | 410 | // ok, we have four consecutive marker bytes; find start |
duke@435 | 411 | u_char* q = p - 4; |
duke@435 | 412 | while (*q == badResourceValue) q--; |
duke@435 | 413 | return q + 1; |
duke@435 | 414 | } |
duke@435 | 415 | |
duke@435 | 416 | u_char* find_cushion_forwards(u_char* start) { |
duke@435 | 417 | u_char* p = start; |
duke@435 | 418 | while (p[0] != badResourceValue || p[1] != badResourceValue || |
duke@435 | 419 | p[2] != badResourceValue || p[3] != badResourceValue) p++; |
duke@435 | 420 | // ok, we have four consecutive marker bytes; find end of cushion |
duke@435 | 421 | u_char* q = p + 4; |
duke@435 | 422 | while (*q == badResourceValue) q++; |
duke@435 | 423 | return q - MallocCushion; |
duke@435 | 424 | } |
duke@435 | 425 | |
duke@435 | 426 | void print_neighbor_blocks(void* ptr) { |
duke@435 | 427 | // find block allocated before ptr (not entirely crash-proof) |
duke@435 | 428 | if (MallocCushion < 4) { |
duke@435 | 429 | tty->print_cr("### cannot find previous block (MallocCushion < 4)"); |
duke@435 | 430 | return; |
duke@435 | 431 | } |
duke@435 | 432 | u_char* start_of_this_block = (u_char*)ptr - space_before; |
duke@435 | 433 | u_char* end_of_prev_block_data = start_of_this_block - space_after -1; |
duke@435 | 434 | // look for cushion in front of prev. block |
duke@435 | 435 | u_char* start_of_prev_block = find_cushion_backwards(end_of_prev_block_data); |
duke@435 | 436 | ptrdiff_t size = *size_addr_from_base(start_of_prev_block); |
duke@435 | 437 | u_char* obj = start_of_prev_block + space_before; |
duke@435 | 438 | if (size <= 0 ) { |
duke@435 | 439 | // start is bad; mayhave been confused by OS data inbetween objects |
duke@435 | 440 | // search one more backwards |
duke@435 | 441 | start_of_prev_block = find_cushion_backwards(start_of_prev_block); |
duke@435 | 442 | size = *size_addr_from_base(start_of_prev_block); |
duke@435 | 443 | obj = start_of_prev_block + space_before; |
duke@435 | 444 | } |
duke@435 | 445 | |
duke@435 | 446 | if (start_of_prev_block + space_before + size + space_after == start_of_this_block) { |
duke@435 | 447 | tty->print_cr("### previous object: %p (%ld bytes)", obj, size); |
duke@435 | 448 | } else { |
duke@435 | 449 | tty->print_cr("### previous object (not sure if correct): %p (%ld bytes)", obj, size); |
duke@435 | 450 | } |
duke@435 | 451 | |
duke@435 | 452 | // now find successor block |
duke@435 | 453 | u_char* start_of_next_block = (u_char*)ptr + *size_addr_from_obj(ptr) + space_after; |
duke@435 | 454 | start_of_next_block = find_cushion_forwards(start_of_next_block); |
duke@435 | 455 | u_char* next_obj = start_of_next_block + space_before; |
duke@435 | 456 | ptrdiff_t next_size = *size_addr_from_base(start_of_next_block); |
duke@435 | 457 | if (start_of_next_block[0] == badResourceValue && |
duke@435 | 458 | start_of_next_block[1] == badResourceValue && |
duke@435 | 459 | start_of_next_block[2] == badResourceValue && |
duke@435 | 460 | start_of_next_block[3] == badResourceValue) { |
duke@435 | 461 | tty->print_cr("### next object: %p (%ld bytes)", next_obj, next_size); |
duke@435 | 462 | } else { |
duke@435 | 463 | tty->print_cr("### next object (not sure if correct): %p (%ld bytes)", next_obj, next_size); |
duke@435 | 464 | } |
duke@435 | 465 | } |
duke@435 | 466 | |
duke@435 | 467 | |
duke@435 | 468 | void report_heap_error(void* memblock, void* bad, const char* where) { |
duke@435 | 469 | tty->print_cr("## nof_mallocs = %d, nof_frees = %d", os::num_mallocs, os::num_frees); |
duke@435 | 470 | tty->print_cr("## memory stomp: byte at %p %s object %p", bad, where, memblock); |
duke@435 | 471 | print_neighbor_blocks(memblock); |
duke@435 | 472 | fatal("memory stomping error"); |
duke@435 | 473 | } |
duke@435 | 474 | |
duke@435 | 475 | void verify_block(void* memblock) { |
duke@435 | 476 | size_t size = get_size(memblock); |
duke@435 | 477 | if (MallocCushion) { |
duke@435 | 478 | u_char* ptr = (u_char*)memblock - space_before; |
duke@435 | 479 | for (int i = 0; i < MallocCushion; i++) { |
duke@435 | 480 | if (ptr[i] != badResourceValue) { |
duke@435 | 481 | report_heap_error(memblock, ptr+i, "in front of"); |
duke@435 | 482 | } |
duke@435 | 483 | } |
duke@435 | 484 | u_char* end = (u_char*)memblock + size + space_after; |
duke@435 | 485 | for (int j = -MallocCushion; j < 0; j++) { |
duke@435 | 486 | if (end[j] != badResourceValue) { |
duke@435 | 487 | report_heap_error(memblock, end+j, "after"); |
duke@435 | 488 | } |
duke@435 | 489 | } |
duke@435 | 490 | } |
duke@435 | 491 | } |
duke@435 | 492 | #endif |
duke@435 | 493 | |
duke@435 | 494 | void* os::malloc(size_t size) { |
duke@435 | 495 | NOT_PRODUCT(num_mallocs++); |
duke@435 | 496 | NOT_PRODUCT(alloc_bytes += size); |
duke@435 | 497 | |
duke@435 | 498 | if (size == 0) { |
duke@435 | 499 | // return a valid pointer if size is zero |
duke@435 | 500 | // if NULL is returned the calling functions assume out of memory. |
duke@435 | 501 | size = 1; |
duke@435 | 502 | } |
duke@435 | 503 | |
duke@435 | 504 | NOT_PRODUCT(if (MallocVerifyInterval > 0) check_heap()); |
duke@435 | 505 | u_char* ptr = (u_char*)::malloc(size + space_before + space_after); |
duke@435 | 506 | #ifdef ASSERT |
duke@435 | 507 | if (ptr == NULL) return NULL; |
duke@435 | 508 | if (MallocCushion) { |
duke@435 | 509 | for (u_char* p = ptr; p < ptr + MallocCushion; p++) *p = (u_char)badResourceValue; |
duke@435 | 510 | u_char* end = ptr + space_before + size; |
duke@435 | 511 | for (u_char* pq = ptr+MallocCushion; pq < end; pq++) *pq = (u_char)uninitBlockPad; |
duke@435 | 512 | for (u_char* q = end; q < end + MallocCushion; q++) *q = (u_char)badResourceValue; |
duke@435 | 513 | } |
duke@435 | 514 | // put size just before data |
duke@435 | 515 | *size_addr_from_base(ptr) = size; |
duke@435 | 516 | #endif |
duke@435 | 517 | u_char* memblock = ptr + space_before; |
duke@435 | 518 | if ((intptr_t)memblock == (intptr_t)MallocCatchPtr) { |
duke@435 | 519 | tty->print_cr("os::malloc caught, %lu bytes --> %p", size, memblock); |
duke@435 | 520 | breakpoint(); |
duke@435 | 521 | } |
duke@435 | 522 | debug_only(if (paranoid) verify_block(memblock)); |
duke@435 | 523 | if (PrintMalloc && tty != NULL) tty->print_cr("os::malloc %lu bytes --> %p", size, memblock); |
duke@435 | 524 | return memblock; |
duke@435 | 525 | } |
duke@435 | 526 | |
duke@435 | 527 | |
duke@435 | 528 | void* os::realloc(void *memblock, size_t size) { |
duke@435 | 529 | NOT_PRODUCT(num_mallocs++); |
duke@435 | 530 | NOT_PRODUCT(alloc_bytes += size); |
duke@435 | 531 | #ifndef ASSERT |
duke@435 | 532 | return ::realloc(memblock, size); |
duke@435 | 533 | #else |
duke@435 | 534 | if (memblock == NULL) { |
duke@435 | 535 | return os::malloc(size); |
duke@435 | 536 | } |
duke@435 | 537 | if ((intptr_t)memblock == (intptr_t)MallocCatchPtr) { |
duke@435 | 538 | tty->print_cr("os::realloc caught %p", memblock); |
duke@435 | 539 | breakpoint(); |
duke@435 | 540 | } |
duke@435 | 541 | verify_block(memblock); |
duke@435 | 542 | NOT_PRODUCT(if (MallocVerifyInterval > 0) check_heap()); |
duke@435 | 543 | if (size == 0) return NULL; |
duke@435 | 544 | // always move the block |
duke@435 | 545 | void* ptr = malloc(size); |
duke@435 | 546 | if (PrintMalloc) tty->print_cr("os::remalloc %lu bytes, %p --> %p", size, memblock, ptr); |
duke@435 | 547 | // Copy to new memory if malloc didn't fail |
duke@435 | 548 | if ( ptr != NULL ) { |
duke@435 | 549 | memcpy(ptr, memblock, MIN2(size, get_size(memblock))); |
duke@435 | 550 | if (paranoid) verify_block(ptr); |
duke@435 | 551 | if ((intptr_t)ptr == (intptr_t)MallocCatchPtr) { |
duke@435 | 552 | tty->print_cr("os::realloc caught, %lu bytes --> %p", size, ptr); |
duke@435 | 553 | breakpoint(); |
duke@435 | 554 | } |
duke@435 | 555 | free(memblock); |
duke@435 | 556 | } |
duke@435 | 557 | return ptr; |
duke@435 | 558 | #endif |
duke@435 | 559 | } |
duke@435 | 560 | |
duke@435 | 561 | |
duke@435 | 562 | void os::free(void *memblock) { |
duke@435 | 563 | NOT_PRODUCT(num_frees++); |
duke@435 | 564 | #ifdef ASSERT |
duke@435 | 565 | if (memblock == NULL) return; |
duke@435 | 566 | if ((intptr_t)memblock == (intptr_t)MallocCatchPtr) { |
duke@435 | 567 | if (tty != NULL) tty->print_cr("os::free caught %p", memblock); |
duke@435 | 568 | breakpoint(); |
duke@435 | 569 | } |
duke@435 | 570 | verify_block(memblock); |
duke@435 | 571 | if (PrintMalloc && tty != NULL) |
duke@435 | 572 | // tty->print_cr("os::free %p", memblock); |
duke@435 | 573 | fprintf(stderr, "os::free %p\n", memblock); |
duke@435 | 574 | NOT_PRODUCT(if (MallocVerifyInterval > 0) check_heap()); |
duke@435 | 575 | // Added by detlefs. |
duke@435 | 576 | if (MallocCushion) { |
duke@435 | 577 | u_char* ptr = (u_char*)memblock - space_before; |
duke@435 | 578 | for (u_char* p = ptr; p < ptr + MallocCushion; p++) { |
duke@435 | 579 | guarantee(*p == badResourceValue, |
duke@435 | 580 | "Thing freed should be malloc result."); |
duke@435 | 581 | *p = (u_char)freeBlockPad; |
duke@435 | 582 | } |
duke@435 | 583 | size_t size = get_size(memblock); |
duke@435 | 584 | u_char* end = ptr + space_before + size; |
duke@435 | 585 | for (u_char* q = end; q < end + MallocCushion; q++) { |
duke@435 | 586 | guarantee(*q == badResourceValue, |
duke@435 | 587 | "Thing freed should be malloc result."); |
duke@435 | 588 | *q = (u_char)freeBlockPad; |
duke@435 | 589 | } |
duke@435 | 590 | } |
duke@435 | 591 | #endif |
duke@435 | 592 | ::free((char*)memblock - space_before); |
duke@435 | 593 | } |
duke@435 | 594 | |
duke@435 | 595 | void os::init_random(long initval) { |
duke@435 | 596 | _rand_seed = initval; |
duke@435 | 597 | } |
duke@435 | 598 | |
duke@435 | 599 | |
duke@435 | 600 | long os::random() { |
duke@435 | 601 | /* standard, well-known linear congruential random generator with |
duke@435 | 602 | * next_rand = (16807*seed) mod (2**31-1) |
duke@435 | 603 | * see |
duke@435 | 604 | * (1) "Random Number Generators: Good Ones Are Hard to Find", |
duke@435 | 605 | * S.K. Park and K.W. Miller, Communications of the ACM 31:10 (Oct 1988), |
duke@435 | 606 | * (2) "Two Fast Implementations of the 'Minimal Standard' Random |
duke@435 | 607 | * Number Generator", David G. Carta, Comm. ACM 33, 1 (Jan 1990), pp. 87-88. |
duke@435 | 608 | */ |
duke@435 | 609 | const long a = 16807; |
duke@435 | 610 | const unsigned long m = 2147483647; |
duke@435 | 611 | const long q = m / a; assert(q == 127773, "weird math"); |
duke@435 | 612 | const long r = m % a; assert(r == 2836, "weird math"); |
duke@435 | 613 | |
duke@435 | 614 | // compute az=2^31p+q |
duke@435 | 615 | unsigned long lo = a * (long)(_rand_seed & 0xFFFF); |
duke@435 | 616 | unsigned long hi = a * (long)((unsigned long)_rand_seed >> 16); |
duke@435 | 617 | lo += (hi & 0x7FFF) << 16; |
duke@435 | 618 | |
duke@435 | 619 | // if q overflowed, ignore the overflow and increment q |
duke@435 | 620 | if (lo > m) { |
duke@435 | 621 | lo &= m; |
duke@435 | 622 | ++lo; |
duke@435 | 623 | } |
duke@435 | 624 | lo += hi >> 15; |
duke@435 | 625 | |
duke@435 | 626 | // if (p+q) overflowed, ignore the overflow and increment (p+q) |
duke@435 | 627 | if (lo > m) { |
duke@435 | 628 | lo &= m; |
duke@435 | 629 | ++lo; |
duke@435 | 630 | } |
duke@435 | 631 | return (_rand_seed = lo); |
duke@435 | 632 | } |
duke@435 | 633 | |
duke@435 | 634 | // The INITIALIZED state is distinguished from the SUSPENDED state because the |
duke@435 | 635 | // conditions in which a thread is first started are different from those in which |
duke@435 | 636 | // a suspension is resumed. These differences make it hard for us to apply the |
duke@435 | 637 | // tougher checks when starting threads that we want to do when resuming them. |
duke@435 | 638 | // However, when start_thread is called as a result of Thread.start, on a Java |
duke@435 | 639 | // thread, the operation is synchronized on the Java Thread object. So there |
duke@435 | 640 | // cannot be a race to start the thread and hence for the thread to exit while |
duke@435 | 641 | // we are working on it. Non-Java threads that start Java threads either have |
duke@435 | 642 | // to do so in a context in which races are impossible, or should do appropriate |
duke@435 | 643 | // locking. |
duke@435 | 644 | |
duke@435 | 645 | void os::start_thread(Thread* thread) { |
duke@435 | 646 | // guard suspend/resume |
duke@435 | 647 | MutexLockerEx ml(thread->SR_lock(), Mutex::_no_safepoint_check_flag); |
duke@435 | 648 | OSThread* osthread = thread->osthread(); |
duke@435 | 649 | osthread->set_state(RUNNABLE); |
duke@435 | 650 | pd_start_thread(thread); |
duke@435 | 651 | } |
duke@435 | 652 | |
duke@435 | 653 | //--------------------------------------------------------------------------- |
duke@435 | 654 | // Helper functions for fatal error handler |
duke@435 | 655 | |
duke@435 | 656 | void os::print_hex_dump(outputStream* st, address start, address end, int unitsize) { |
duke@435 | 657 | assert(unitsize == 1 || unitsize == 2 || unitsize == 4 || unitsize == 8, "just checking"); |
duke@435 | 658 | |
duke@435 | 659 | int cols = 0; |
duke@435 | 660 | int cols_per_line = 0; |
duke@435 | 661 | switch (unitsize) { |
duke@435 | 662 | case 1: cols_per_line = 16; break; |
duke@435 | 663 | case 2: cols_per_line = 8; break; |
duke@435 | 664 | case 4: cols_per_line = 4; break; |
duke@435 | 665 | case 8: cols_per_line = 2; break; |
duke@435 | 666 | default: return; |
duke@435 | 667 | } |
duke@435 | 668 | |
duke@435 | 669 | address p = start; |
duke@435 | 670 | st->print(PTR_FORMAT ": ", start); |
duke@435 | 671 | while (p < end) { |
duke@435 | 672 | switch (unitsize) { |
duke@435 | 673 | case 1: st->print("%02x", *(u1*)p); break; |
duke@435 | 674 | case 2: st->print("%04x", *(u2*)p); break; |
duke@435 | 675 | case 4: st->print("%08x", *(u4*)p); break; |
duke@435 | 676 | case 8: st->print("%016" FORMAT64_MODIFIER "x", *(u8*)p); break; |
duke@435 | 677 | } |
duke@435 | 678 | p += unitsize; |
duke@435 | 679 | cols++; |
duke@435 | 680 | if (cols >= cols_per_line && p < end) { |
duke@435 | 681 | cols = 0; |
duke@435 | 682 | st->cr(); |
duke@435 | 683 | st->print(PTR_FORMAT ": ", p); |
duke@435 | 684 | } else { |
duke@435 | 685 | st->print(" "); |
duke@435 | 686 | } |
duke@435 | 687 | } |
duke@435 | 688 | st->cr(); |
duke@435 | 689 | } |
duke@435 | 690 | |
duke@435 | 691 | void os::print_environment_variables(outputStream* st, const char** env_list, |
duke@435 | 692 | char* buffer, int len) { |
duke@435 | 693 | if (env_list) { |
duke@435 | 694 | st->print_cr("Environment Variables:"); |
duke@435 | 695 | |
duke@435 | 696 | for (int i = 0; env_list[i] != NULL; i++) { |
duke@435 | 697 | if (getenv(env_list[i], buffer, len)) { |
duke@435 | 698 | st->print(env_list[i]); |
duke@435 | 699 | st->print("="); |
duke@435 | 700 | st->print_cr(buffer); |
duke@435 | 701 | } |
duke@435 | 702 | } |
duke@435 | 703 | } |
duke@435 | 704 | } |
duke@435 | 705 | |
duke@435 | 706 | void os::print_cpu_info(outputStream* st) { |
duke@435 | 707 | // cpu |
duke@435 | 708 | st->print("CPU:"); |
duke@435 | 709 | st->print("total %d", os::processor_count()); |
duke@435 | 710 | // It's not safe to query number of active processors after crash |
duke@435 | 711 | // st->print("(active %d)", os::active_processor_count()); |
duke@435 | 712 | st->print(" %s", VM_Version::cpu_features()); |
duke@435 | 713 | st->cr(); |
duke@435 | 714 | } |
duke@435 | 715 | |
duke@435 | 716 | void os::print_date_and_time(outputStream *st) { |
duke@435 | 717 | time_t tloc; |
duke@435 | 718 | (void)time(&tloc); |
duke@435 | 719 | st->print("time: %s", ctime(&tloc)); // ctime adds newline. |
duke@435 | 720 | |
duke@435 | 721 | double t = os::elapsedTime(); |
duke@435 | 722 | // NOTE: It tends to crash after a SEGV if we want to printf("%f",...) in |
duke@435 | 723 | // Linux. Must be a bug in glibc ? Workaround is to round "t" to int |
duke@435 | 724 | // before printf. We lost some precision, but who cares? |
duke@435 | 725 | st->print_cr("elapsed time: %d seconds", (int)t); |
duke@435 | 726 | } |
duke@435 | 727 | |
duke@435 | 728 | |
duke@435 | 729 | // Looks like all platforms except IA64 can use the same function to check |
duke@435 | 730 | // if C stack is walkable beyond current frame. The check for fp() is not |
duke@435 | 731 | // necessary on Sparc, but it's harmless. |
duke@435 | 732 | bool os::is_first_C_frame(frame* fr) { |
duke@435 | 733 | #ifdef IA64 |
duke@435 | 734 | // In order to walk native frames on Itanium, we need to access the unwind |
duke@435 | 735 | // table, which is inside ELF. We don't want to parse ELF after fatal error, |
duke@435 | 736 | // so return true for IA64. If we need to support C stack walking on IA64, |
duke@435 | 737 | // this function needs to be moved to CPU specific files, as fp() on IA64 |
duke@435 | 738 | // is register stack, which grows towards higher memory address. |
duke@435 | 739 | return true; |
duke@435 | 740 | #endif |
duke@435 | 741 | |
duke@435 | 742 | // Load up sp, fp, sender sp and sender fp, check for reasonable values. |
duke@435 | 743 | // Check usp first, because if that's bad the other accessors may fault |
duke@435 | 744 | // on some architectures. Ditto ufp second, etc. |
duke@435 | 745 | uintptr_t fp_align_mask = (uintptr_t)(sizeof(address)-1); |
duke@435 | 746 | // sp on amd can be 32 bit aligned. |
duke@435 | 747 | uintptr_t sp_align_mask = (uintptr_t)(sizeof(int)-1); |
duke@435 | 748 | |
duke@435 | 749 | uintptr_t usp = (uintptr_t)fr->sp(); |
duke@435 | 750 | if ((usp & sp_align_mask) != 0) return true; |
duke@435 | 751 | |
duke@435 | 752 | uintptr_t ufp = (uintptr_t)fr->fp(); |
duke@435 | 753 | if ((ufp & fp_align_mask) != 0) return true; |
duke@435 | 754 | |
duke@435 | 755 | uintptr_t old_sp = (uintptr_t)fr->sender_sp(); |
duke@435 | 756 | if ((old_sp & sp_align_mask) != 0) return true; |
duke@435 | 757 | if (old_sp == 0 || old_sp == (uintptr_t)-1) return true; |
duke@435 | 758 | |
duke@435 | 759 | uintptr_t old_fp = (uintptr_t)fr->link(); |
duke@435 | 760 | if ((old_fp & fp_align_mask) != 0) return true; |
duke@435 | 761 | if (old_fp == 0 || old_fp == (uintptr_t)-1 || old_fp == ufp) return true; |
duke@435 | 762 | |
duke@435 | 763 | // stack grows downwards; if old_fp is below current fp or if the stack |
duke@435 | 764 | // frame is too large, either the stack is corrupted or fp is not saved |
duke@435 | 765 | // on stack (i.e. on x86, ebp may be used as general register). The stack |
duke@435 | 766 | // is not walkable beyond current frame. |
duke@435 | 767 | if (old_fp < ufp) return true; |
duke@435 | 768 | if (old_fp - ufp > 64 * K) return true; |
duke@435 | 769 | |
duke@435 | 770 | return false; |
duke@435 | 771 | } |
duke@435 | 772 | |
duke@435 | 773 | #ifdef ASSERT |
duke@435 | 774 | extern "C" void test_random() { |
duke@435 | 775 | const double m = 2147483647; |
duke@435 | 776 | double mean = 0.0, variance = 0.0, t; |
duke@435 | 777 | long reps = 10000; |
duke@435 | 778 | unsigned long seed = 1; |
duke@435 | 779 | |
duke@435 | 780 | tty->print_cr("seed %ld for %ld repeats...", seed, reps); |
duke@435 | 781 | os::init_random(seed); |
duke@435 | 782 | long num; |
duke@435 | 783 | for (int k = 0; k < reps; k++) { |
duke@435 | 784 | num = os::random(); |
duke@435 | 785 | double u = (double)num / m; |
duke@435 | 786 | assert(u >= 0.0 && u <= 1.0, "bad random number!"); |
duke@435 | 787 | |
duke@435 | 788 | // calculate mean and variance of the random sequence |
duke@435 | 789 | mean += u; |
duke@435 | 790 | variance += (u*u); |
duke@435 | 791 | } |
duke@435 | 792 | mean /= reps; |
duke@435 | 793 | variance /= (reps - 1); |
duke@435 | 794 | |
duke@435 | 795 | assert(num == 1043618065, "bad seed"); |
duke@435 | 796 | tty->print_cr("mean of the 1st 10000 numbers: %f", mean); |
duke@435 | 797 | tty->print_cr("variance of the 1st 10000 numbers: %f", variance); |
duke@435 | 798 | const double eps = 0.0001; |
duke@435 | 799 | t = fabsd(mean - 0.5018); |
duke@435 | 800 | assert(t < eps, "bad mean"); |
duke@435 | 801 | t = (variance - 0.3355) < 0.0 ? -(variance - 0.3355) : variance - 0.3355; |
duke@435 | 802 | assert(t < eps, "bad variance"); |
duke@435 | 803 | } |
duke@435 | 804 | #endif |
duke@435 | 805 | |
duke@435 | 806 | |
duke@435 | 807 | // Set up the boot classpath. |
duke@435 | 808 | |
duke@435 | 809 | char* os::format_boot_path(const char* format_string, |
duke@435 | 810 | const char* home, |
duke@435 | 811 | int home_len, |
duke@435 | 812 | char fileSep, |
duke@435 | 813 | char pathSep) { |
duke@435 | 814 | assert((fileSep == '/' && pathSep == ':') || |
duke@435 | 815 | (fileSep == '\\' && pathSep == ';'), "unexpected seperator chars"); |
duke@435 | 816 | |
duke@435 | 817 | // Scan the format string to determine the length of the actual |
duke@435 | 818 | // boot classpath, and handle platform dependencies as well. |
duke@435 | 819 | int formatted_path_len = 0; |
duke@435 | 820 | const char* p; |
duke@435 | 821 | for (p = format_string; *p != 0; ++p) { |
duke@435 | 822 | if (*p == '%') formatted_path_len += home_len - 1; |
duke@435 | 823 | ++formatted_path_len; |
duke@435 | 824 | } |
duke@435 | 825 | |
duke@435 | 826 | char* formatted_path = NEW_C_HEAP_ARRAY(char, formatted_path_len + 1); |
duke@435 | 827 | if (formatted_path == NULL) { |
duke@435 | 828 | return NULL; |
duke@435 | 829 | } |
duke@435 | 830 | |
duke@435 | 831 | // Create boot classpath from format, substituting separator chars and |
duke@435 | 832 | // java home directory. |
duke@435 | 833 | char* q = formatted_path; |
duke@435 | 834 | for (p = format_string; *p != 0; ++p) { |
duke@435 | 835 | switch (*p) { |
duke@435 | 836 | case '%': |
duke@435 | 837 | strcpy(q, home); |
duke@435 | 838 | q += home_len; |
duke@435 | 839 | break; |
duke@435 | 840 | case '/': |
duke@435 | 841 | *q++ = fileSep; |
duke@435 | 842 | break; |
duke@435 | 843 | case ':': |
duke@435 | 844 | *q++ = pathSep; |
duke@435 | 845 | break; |
duke@435 | 846 | default: |
duke@435 | 847 | *q++ = *p; |
duke@435 | 848 | } |
duke@435 | 849 | } |
duke@435 | 850 | *q = '\0'; |
duke@435 | 851 | |
duke@435 | 852 | assert((q - formatted_path) == formatted_path_len, "formatted_path size botched"); |
duke@435 | 853 | return formatted_path; |
duke@435 | 854 | } |
duke@435 | 855 | |
duke@435 | 856 | |
duke@435 | 857 | bool os::set_boot_path(char fileSep, char pathSep) { |
duke@435 | 858 | |
duke@435 | 859 | const char* home = Arguments::get_java_home(); |
duke@435 | 860 | int home_len = (int)strlen(home); |
duke@435 | 861 | |
duke@435 | 862 | static const char* meta_index_dir_format = "%/lib/"; |
duke@435 | 863 | static const char* meta_index_format = "%/lib/meta-index"; |
duke@435 | 864 | char* meta_index = format_boot_path(meta_index_format, home, home_len, fileSep, pathSep); |
duke@435 | 865 | if (meta_index == NULL) return false; |
duke@435 | 866 | char* meta_index_dir = format_boot_path(meta_index_dir_format, home, home_len, fileSep, pathSep); |
duke@435 | 867 | if (meta_index_dir == NULL) return false; |
duke@435 | 868 | Arguments::set_meta_index_path(meta_index, meta_index_dir); |
duke@435 | 869 | |
duke@435 | 870 | // Any modification to the JAR-file list, for the boot classpath must be |
duke@435 | 871 | // aligned with install/install/make/common/Pack.gmk. Note: boot class |
duke@435 | 872 | // path class JARs, are stripped for StackMapTable to reduce download size. |
duke@435 | 873 | static const char classpath_format[] = |
duke@435 | 874 | "%/lib/resources.jar:" |
duke@435 | 875 | "%/lib/rt.jar:" |
duke@435 | 876 | "%/lib/sunrsasign.jar:" |
duke@435 | 877 | "%/lib/jsse.jar:" |
duke@435 | 878 | "%/lib/jce.jar:" |
duke@435 | 879 | "%/lib/charsets.jar:" |
duke@435 | 880 | "%/classes"; |
duke@435 | 881 | char* sysclasspath = format_boot_path(classpath_format, home, home_len, fileSep, pathSep); |
duke@435 | 882 | if (sysclasspath == NULL) return false; |
duke@435 | 883 | Arguments::set_sysclasspath(sysclasspath); |
duke@435 | 884 | |
duke@435 | 885 | return true; |
duke@435 | 886 | } |
duke@435 | 887 | |
duke@435 | 888 | void os::set_memory_serialize_page(address page) { |
duke@435 | 889 | int count = log2_intptr(sizeof(class JavaThread)) - log2_intptr(64); |
duke@435 | 890 | _mem_serialize_page = (volatile int32_t *)page; |
duke@435 | 891 | // We initialize the serialization page shift count here |
duke@435 | 892 | // We assume a cache line size of 64 bytes |
duke@435 | 893 | assert(SerializePageShiftCount == count, |
duke@435 | 894 | "thread size changed, fix SerializePageShiftCount constant"); |
duke@435 | 895 | set_serialize_page_mask((uintptr_t)(vm_page_size() - sizeof(int32_t))); |
duke@435 | 896 | } |
duke@435 | 897 | |
xlu@490 | 898 | static volatile intptr_t SerializePageLock = 0; |
xlu@490 | 899 | |
duke@435 | 900 | // This method is called from signal handler when SIGSEGV occurs while the current |
duke@435 | 901 | // thread tries to store to the "read-only" memory serialize page during state |
duke@435 | 902 | // transition. |
duke@435 | 903 | void os::block_on_serialize_page_trap() { |
duke@435 | 904 | if (TraceSafepoint) { |
duke@435 | 905 | tty->print_cr("Block until the serialize page permission restored"); |
duke@435 | 906 | } |
xlu@490 | 907 | // When VMThread is holding the SerializePageLock during modifying the |
duke@435 | 908 | // access permission of the memory serialize page, the following call |
duke@435 | 909 | // will block until the permission of that page is restored to rw. |
duke@435 | 910 | // Generally, it is unsafe to manipulate locks in signal handlers, but in |
duke@435 | 911 | // this case, it's OK as the signal is synchronous and we know precisely when |
xlu@490 | 912 | // it can occur. |
xlu@490 | 913 | Thread::muxAcquire(&SerializePageLock, "set_memory_serialize_page"); |
xlu@490 | 914 | Thread::muxRelease(&SerializePageLock); |
duke@435 | 915 | } |
duke@435 | 916 | |
duke@435 | 917 | // Serialize all thread state variables |
duke@435 | 918 | void os::serialize_thread_states() { |
duke@435 | 919 | // On some platforms such as Solaris & Linux, the time duration of the page |
duke@435 | 920 | // permission restoration is observed to be much longer than expected due to |
duke@435 | 921 | // scheduler starvation problem etc. To avoid the long synchronization |
xlu@490 | 922 | // time and expensive page trap spinning, 'SerializePageLock' is used to block |
xlu@490 | 923 | // the mutator thread if such case is encountered. See bug 6546278 for details. |
xlu@490 | 924 | Thread::muxAcquire(&SerializePageLock, "serialize_thread_states"); |
duke@435 | 925 | os::protect_memory( (char *)os::get_memory_serialize_page(), os::vm_page_size() ); |
duke@435 | 926 | os::unguard_memory( (char *)os::get_memory_serialize_page(), os::vm_page_size() ); |
xlu@490 | 927 | Thread::muxRelease(&SerializePageLock); |
duke@435 | 928 | } |
duke@435 | 929 | |
duke@435 | 930 | // Returns true if the current stack pointer is above the stack shadow |
duke@435 | 931 | // pages, false otherwise. |
duke@435 | 932 | |
duke@435 | 933 | bool os::stack_shadow_pages_available(Thread *thread, methodHandle method) { |
duke@435 | 934 | assert(StackRedPages > 0 && StackYellowPages > 0,"Sanity check"); |
duke@435 | 935 | address sp = current_stack_pointer(); |
duke@435 | 936 | // Check if we have StackShadowPages above the yellow zone. This parameter |
duke@435 | 937 | // is dependant on the depth of the maximum VM call stack possible from |
duke@435 | 938 | // the handler for stack overflow. 'instanceof' in the stack overflow |
duke@435 | 939 | // handler or a println uses at least 8k stack of VM and native code |
duke@435 | 940 | // respectively. |
duke@435 | 941 | const int framesize_in_bytes = |
duke@435 | 942 | Interpreter::size_top_interpreter_activation(method()) * wordSize; |
duke@435 | 943 | int reserved_area = ((StackShadowPages + StackRedPages + StackYellowPages) |
duke@435 | 944 | * vm_page_size()) + framesize_in_bytes; |
duke@435 | 945 | // The very lower end of the stack |
duke@435 | 946 | address stack_limit = thread->stack_base() - thread->stack_size(); |
duke@435 | 947 | return (sp > (stack_limit + reserved_area)); |
duke@435 | 948 | } |
duke@435 | 949 | |
duke@435 | 950 | size_t os::page_size_for_region(size_t region_min_size, size_t region_max_size, |
duke@435 | 951 | uint min_pages) |
duke@435 | 952 | { |
duke@435 | 953 | assert(min_pages > 0, "sanity"); |
duke@435 | 954 | if (UseLargePages) { |
duke@435 | 955 | const size_t max_page_size = region_max_size / min_pages; |
duke@435 | 956 | |
duke@435 | 957 | for (unsigned int i = 0; _page_sizes[i] != 0; ++i) { |
duke@435 | 958 | const size_t sz = _page_sizes[i]; |
duke@435 | 959 | const size_t mask = sz - 1; |
duke@435 | 960 | if ((region_min_size & mask) == 0 && (region_max_size & mask) == 0) { |
duke@435 | 961 | // The largest page size with no fragmentation. |
duke@435 | 962 | return sz; |
duke@435 | 963 | } |
duke@435 | 964 | |
duke@435 | 965 | if (sz <= max_page_size) { |
duke@435 | 966 | // The largest page size that satisfies the min_pages requirement. |
duke@435 | 967 | return sz; |
duke@435 | 968 | } |
duke@435 | 969 | } |
duke@435 | 970 | } |
duke@435 | 971 | |
duke@435 | 972 | return vm_page_size(); |
duke@435 | 973 | } |
duke@435 | 974 | |
duke@435 | 975 | #ifndef PRODUCT |
duke@435 | 976 | void os::trace_page_sizes(const char* str, const size_t region_min_size, |
duke@435 | 977 | const size_t region_max_size, const size_t page_size, |
duke@435 | 978 | const char* base, const size_t size) |
duke@435 | 979 | { |
duke@435 | 980 | if (TracePageSizes) { |
duke@435 | 981 | tty->print_cr("%s: min=" SIZE_FORMAT " max=" SIZE_FORMAT |
duke@435 | 982 | " pg_sz=" SIZE_FORMAT " base=" PTR_FORMAT |
duke@435 | 983 | " size=" SIZE_FORMAT, |
duke@435 | 984 | str, region_min_size, region_max_size, |
duke@435 | 985 | page_size, base, size); |
duke@435 | 986 | } |
duke@435 | 987 | } |
duke@435 | 988 | #endif // #ifndef PRODUCT |
duke@435 | 989 | |
duke@435 | 990 | // This is the working definition of a server class machine: |
duke@435 | 991 | // >= 2 physical CPU's and >=2GB of memory, with some fuzz |
duke@435 | 992 | // because the graphics memory (?) sometimes masks physical memory. |
duke@435 | 993 | // If you want to change the definition of a server class machine |
duke@435 | 994 | // on some OS or platform, e.g., >=4GB on Windohs platforms, |
duke@435 | 995 | // then you'll have to parameterize this method based on that state, |
duke@435 | 996 | // as was done for logical processors here, or replicate and |
duke@435 | 997 | // specialize this method for each platform. (Or fix os to have |
duke@435 | 998 | // some inheritance structure and use subclassing. Sigh.) |
duke@435 | 999 | // If you want some platform to always or never behave as a server |
duke@435 | 1000 | // class machine, change the setting of AlwaysActAsServerClassMachine |
duke@435 | 1001 | // and NeverActAsServerClassMachine in globals*.hpp. |
duke@435 | 1002 | bool os::is_server_class_machine() { |
duke@435 | 1003 | // First check for the early returns |
duke@435 | 1004 | if (NeverActAsServerClassMachine) { |
duke@435 | 1005 | return false; |
duke@435 | 1006 | } |
duke@435 | 1007 | if (AlwaysActAsServerClassMachine) { |
duke@435 | 1008 | return true; |
duke@435 | 1009 | } |
duke@435 | 1010 | // Then actually look at the machine |
duke@435 | 1011 | bool result = false; |
duke@435 | 1012 | const unsigned int server_processors = 2; |
duke@435 | 1013 | const julong server_memory = 2UL * G; |
duke@435 | 1014 | // We seem not to get our full complement of memory. |
duke@435 | 1015 | // We allow some part (1/8?) of the memory to be "missing", |
duke@435 | 1016 | // based on the sizes of DIMMs, and maybe graphics cards. |
duke@435 | 1017 | const julong missing_memory = 256UL * M; |
duke@435 | 1018 | |
duke@435 | 1019 | /* Is this a server class machine? */ |
duke@435 | 1020 | if ((os::active_processor_count() >= (int)server_processors) && |
duke@435 | 1021 | (os::physical_memory() >= (server_memory - missing_memory))) { |
duke@435 | 1022 | const unsigned int logical_processors = |
duke@435 | 1023 | VM_Version::logical_processors_per_package(); |
duke@435 | 1024 | if (logical_processors > 1) { |
duke@435 | 1025 | const unsigned int physical_packages = |
duke@435 | 1026 | os::active_processor_count() / logical_processors; |
duke@435 | 1027 | if (physical_packages > server_processors) { |
duke@435 | 1028 | result = true; |
duke@435 | 1029 | } |
duke@435 | 1030 | } else { |
duke@435 | 1031 | result = true; |
duke@435 | 1032 | } |
duke@435 | 1033 | } |
duke@435 | 1034 | return result; |
duke@435 | 1035 | } |