duke@435: /* xdono@631: * Copyright 2003-2008 Sun Microsystems, Inc. All Rights Reserved. duke@435: * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. duke@435: * duke@435: * This code is free software; you can redistribute it and/or modify it duke@435: * under the terms of the GNU General Public License version 2 only, as duke@435: * published by the Free Software Foundation. duke@435: * duke@435: * This code is distributed in the hope that it will be useful, but WITHOUT duke@435: * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or duke@435: * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License duke@435: * version 2 for more details (a copy is included in the LICENSE file that duke@435: * accompanied this code). duke@435: * duke@435: * You should have received a copy of the GNU General Public License version duke@435: * 2 along with this work; if not, write to the Free Software Foundation, duke@435: * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. duke@435: * duke@435: * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara, duke@435: * CA 95054 USA or visit www.sun.com if you need additional information or duke@435: * have any questions. duke@435: * duke@435: */ duke@435: duke@435: # include "incls/_precompiled.incl" duke@435: # include "incls/_jvmtiTagMap.cpp.incl" duke@435: duke@435: // JvmtiTagHashmapEntry duke@435: // duke@435: // Each entry encapsulates a JNI weak reference to the tagged object duke@435: // and the tag value. In addition an entry includes a next pointer which duke@435: // is used to chain entries together. duke@435: duke@435: class JvmtiTagHashmapEntry : public CHeapObj { duke@435: private: duke@435: friend class JvmtiTagMap; duke@435: duke@435: jweak _object; // JNI weak ref to tagged object duke@435: jlong _tag; // the tag duke@435: JvmtiTagHashmapEntry* _next; // next on the list duke@435: duke@435: inline void init(jweak object, jlong tag) { duke@435: _object = object; duke@435: _tag = tag; duke@435: _next = NULL; duke@435: } duke@435: duke@435: // constructor duke@435: JvmtiTagHashmapEntry(jweak object, jlong tag) { init(object, tag); } duke@435: duke@435: public: duke@435: duke@435: // accessor methods duke@435: inline jweak object() const { return _object; } duke@435: inline jlong tag() const { return _tag; } duke@435: duke@435: inline void set_tag(jlong tag) { duke@435: assert(tag != 0, "can't be zero"); duke@435: _tag = tag; duke@435: } duke@435: duke@435: inline JvmtiTagHashmapEntry* next() const { return _next; } duke@435: inline void set_next(JvmtiTagHashmapEntry* next) { _next = next; } duke@435: }; duke@435: duke@435: duke@435: // JvmtiTagHashmap duke@435: // duke@435: // A hashmap is essentially a table of pointers to entries. Entries duke@435: // are hashed to a location, or position in the table, and then duke@435: // chained from that location. The "key" for hashing is address of duke@435: // the object, or oop. The "value" is the JNI weak reference to the duke@435: // object and the tag value. Keys are not stored with the entry. duke@435: // Instead the weak reference is resolved to obtain the key. duke@435: // duke@435: // A hashmap maintains a count of the number entries in the hashmap duke@435: // and resizes if the number of entries exceeds a given threshold. duke@435: // The threshold is specified as a percentage of the size - for duke@435: // example a threshold of 0.75 will trigger the hashmap to resize duke@435: // if the number of entries is >75% of table size. duke@435: // duke@435: // A hashmap provides functions for adding, removing, and finding duke@435: // entries. It also provides a function to iterate over all entries duke@435: // in the hashmap. duke@435: duke@435: class JvmtiTagHashmap : public CHeapObj { duke@435: private: duke@435: friend class JvmtiTagMap; duke@435: duke@435: enum { duke@435: small_trace_threshold = 10000, // threshold for tracing duke@435: medium_trace_threshold = 100000, duke@435: large_trace_threshold = 1000000, duke@435: initial_trace_threshold = small_trace_threshold duke@435: }; duke@435: duke@435: static int _sizes[]; // array of possible hashmap sizes duke@435: int _size; // actual size of the table duke@435: int _size_index; // index into size table duke@435: duke@435: int _entry_count; // number of entries in the hashmap duke@435: duke@435: float _load_factor; // load factor as a % of the size duke@435: int _resize_threshold; // computed threshold to trigger resizing. duke@435: bool _resizing_enabled; // indicates if hashmap can resize duke@435: duke@435: int _trace_threshold; // threshold for trace messages duke@435: duke@435: JvmtiTagHashmapEntry** _table; // the table of entries. duke@435: duke@435: // private accessors duke@435: int resize_threshold() const { return _resize_threshold; } duke@435: int trace_threshold() const { return _trace_threshold; } duke@435: duke@435: // initialize the hashmap duke@435: void init(int size_index=0, float load_factor=4.0f) { duke@435: int initial_size = _sizes[size_index]; duke@435: _size_index = size_index; duke@435: _size = initial_size; duke@435: _entry_count = 0; duke@435: if (TraceJVMTIObjectTagging) { duke@435: _trace_threshold = initial_trace_threshold; duke@435: } else { duke@435: _trace_threshold = -1; duke@435: } duke@435: _load_factor = load_factor; duke@435: _resize_threshold = (int)(_load_factor * _size); duke@435: _resizing_enabled = true; duke@435: size_t s = initial_size * sizeof(JvmtiTagHashmapEntry*); duke@435: _table = (JvmtiTagHashmapEntry**)os::malloc(s); duke@435: if (_table == NULL) { duke@435: vm_exit_out_of_memory(s, "unable to allocate initial hashtable for jvmti object tags"); duke@435: } duke@435: for (int i=0; i> 3) % size; duke@435: #else duke@435: return (addr >> 2) % size; duke@435: #endif duke@435: } duke@435: duke@435: // hash a given key (oop) duke@435: unsigned int hash(oop key) { duke@435: return hash(key, _size); duke@435: } duke@435: duke@435: // resize the hashmap - allocates a large table and re-hashes duke@435: // all entries into the new table. duke@435: void resize() { duke@435: int new_size_index = _size_index+1; duke@435: int new_size = _sizes[new_size_index]; duke@435: if (new_size < 0) { duke@435: // hashmap already at maximum capacity duke@435: return; duke@435: } duke@435: duke@435: // allocate new table duke@435: size_t s = new_size * sizeof(JvmtiTagHashmapEntry*); duke@435: JvmtiTagHashmapEntry** new_table = (JvmtiTagHashmapEntry**)os::malloc(s); duke@435: if (new_table == NULL) { duke@435: warning("unable to allocate larger hashtable for jvmti object tags"); duke@435: set_resizing_enabled(false); duke@435: return; duke@435: } duke@435: duke@435: // initialize new table duke@435: int i; duke@435: for (i=0; inext(); duke@435: oop key = JNIHandles::resolve(entry->object()); duke@435: assert(key != NULL, "jni weak reference cleared!!"); duke@435: unsigned int h = hash(key, new_size); duke@435: JvmtiTagHashmapEntry* anchor = new_table[h]; duke@435: if (anchor == NULL) { duke@435: new_table[h] = entry; duke@435: entry->set_next(NULL); duke@435: } else { duke@435: entry->set_next(anchor); duke@435: new_table[h] = entry; duke@435: } duke@435: entry = next; duke@435: } duke@435: } duke@435: duke@435: // free old table and update settings. duke@435: os::free((void*)_table); duke@435: _table = new_table; duke@435: _size_index = new_size_index; duke@435: _size = new_size; duke@435: duke@435: // compute new resize threshold duke@435: _resize_threshold = (int)(_load_factor * _size); duke@435: } duke@435: duke@435: duke@435: // internal remove function - remove an entry at a given position in the duke@435: // table. duke@435: inline void remove(JvmtiTagHashmapEntry* prev, int pos, JvmtiTagHashmapEntry* entry) { duke@435: assert(pos >= 0 && pos < _size, "out of range"); duke@435: if (prev == NULL) { duke@435: _table[pos] = entry->next(); duke@435: } else { duke@435: prev->set_next(entry->next()); duke@435: } duke@435: assert(_entry_count > 0, "checking"); duke@435: _entry_count--; duke@435: } duke@435: duke@435: // resizing switch duke@435: bool is_resizing_enabled() const { return _resizing_enabled; } duke@435: void set_resizing_enabled(bool enable) { _resizing_enabled = enable; } duke@435: duke@435: // debugging duke@435: void print_memory_usage(); duke@435: void compute_next_trace_threshold(); duke@435: duke@435: public: duke@435: duke@435: // create a JvmtiTagHashmap of a preferred size and optionally a load factor. duke@435: // The preferred size is rounded down to an actual size. duke@435: JvmtiTagHashmap(int size, float load_factor=0.0f) { duke@435: int i=0; duke@435: while (_sizes[i] < size) { duke@435: if (_sizes[i] < 0) { duke@435: assert(i > 0, "sanity check"); duke@435: i--; duke@435: break; duke@435: } duke@435: i++; duke@435: } duke@435: duke@435: // if a load factor is specified then use it, otherwise use default duke@435: if (load_factor > 0.01f) { duke@435: init(i, load_factor); duke@435: } else { duke@435: init(i); duke@435: } duke@435: } duke@435: duke@435: // create a JvmtiTagHashmap with default settings duke@435: JvmtiTagHashmap() { duke@435: init(); duke@435: } duke@435: duke@435: // release table when JvmtiTagHashmap destroyed duke@435: ~JvmtiTagHashmap() { duke@435: if (_table != NULL) { duke@435: os::free((void*)_table); duke@435: _table = NULL; duke@435: } duke@435: } duke@435: duke@435: // accessors duke@435: int size() const { return _size; } duke@435: JvmtiTagHashmapEntry** table() const { return _table; } duke@435: int entry_count() const { return _entry_count; } duke@435: duke@435: // find an entry in the hashmap, returns NULL if not found. duke@435: inline JvmtiTagHashmapEntry* find(oop key) { duke@435: unsigned int h = hash(key); duke@435: JvmtiTagHashmapEntry* entry = _table[h]; duke@435: while (entry != NULL) { duke@435: oop orig_key = JNIHandles::resolve(entry->object()); duke@435: assert(orig_key != NULL, "jni weak reference cleared!!"); duke@435: if (key == orig_key) { duke@435: break; duke@435: } duke@435: entry = entry->next(); duke@435: } duke@435: return entry; duke@435: } duke@435: duke@435: duke@435: // add a new entry to hashmap duke@435: inline void add(oop key, JvmtiTagHashmapEntry* entry) { duke@435: assert(key != NULL, "checking"); duke@435: assert(find(key) == NULL, "duplicate detected"); duke@435: unsigned int h = hash(key); duke@435: JvmtiTagHashmapEntry* anchor = _table[h]; duke@435: if (anchor == NULL) { duke@435: _table[h] = entry; duke@435: entry->set_next(NULL); duke@435: } else { duke@435: entry->set_next(anchor); duke@435: _table[h] = entry; duke@435: } duke@435: duke@435: _entry_count++; duke@435: if (trace_threshold() > 0 && entry_count() >= trace_threshold()) { duke@435: assert(TraceJVMTIObjectTagging, "should only get here when tracing"); duke@435: print_memory_usage(); duke@435: compute_next_trace_threshold(); duke@435: } duke@435: duke@435: // if the number of entries exceed the threshold then resize duke@435: if (entry_count() > resize_threshold() && is_resizing_enabled()) { duke@435: resize(); duke@435: } duke@435: } duke@435: duke@435: // remove an entry with the given key. duke@435: inline JvmtiTagHashmapEntry* remove(oop key) { duke@435: unsigned int h = hash(key); duke@435: JvmtiTagHashmapEntry* entry = _table[h]; duke@435: JvmtiTagHashmapEntry* prev = NULL; duke@435: while (entry != NULL) { duke@435: oop orig_key = JNIHandles::resolve(entry->object()); duke@435: assert(orig_key != NULL, "jni weak reference cleared!!"); duke@435: if (key == orig_key) { duke@435: break; duke@435: } duke@435: prev = entry; duke@435: entry = entry->next(); duke@435: } duke@435: if (entry != NULL) { duke@435: remove(prev, h, entry); duke@435: } duke@435: return entry; duke@435: } duke@435: duke@435: // iterate over all entries in the hashmap duke@435: void entry_iterate(JvmtiTagHashmapEntryClosure* closure); duke@435: }; duke@435: duke@435: // possible hashmap sizes - odd primes that roughly double in size. duke@435: // To avoid excessive resizing the odd primes from 4801-76831 and duke@435: // 76831-307261 have been removed. The list must be terminated by -1. duke@435: int JvmtiTagHashmap::_sizes[] = { 4801, 76831, 307261, 614563, 1228891, duke@435: 2457733, 4915219, 9830479, 19660831, 39321619, 78643219, -1 }; duke@435: duke@435: duke@435: // A supporting class for iterating over all entries in Hashmap duke@435: class JvmtiTagHashmapEntryClosure { duke@435: public: duke@435: virtual void do_entry(JvmtiTagHashmapEntry* entry) = 0; duke@435: }; duke@435: duke@435: duke@435: // iterate over all entries in the hashmap duke@435: void JvmtiTagHashmap::entry_iterate(JvmtiTagHashmapEntryClosure* closure) { duke@435: for (int i=0; i<_size; i++) { duke@435: JvmtiTagHashmapEntry* entry = _table[i]; duke@435: JvmtiTagHashmapEntry* prev = NULL; duke@435: while (entry != NULL) { duke@435: // obtain the next entry before invoking do_entry - this is duke@435: // necessary because do_entry may remove the entry from the duke@435: // hashmap. duke@435: JvmtiTagHashmapEntry* next = entry->next(); duke@435: closure->do_entry(entry); duke@435: entry = next; duke@435: } duke@435: } duke@435: } duke@435: duke@435: // debugging duke@435: void JvmtiTagHashmap::print_memory_usage() { duke@435: intptr_t p = (intptr_t)this; duke@435: tty->print("[JvmtiTagHashmap @ " INTPTR_FORMAT, p); duke@435: duke@435: // table + entries in KB duke@435: int hashmap_usage = (size()*sizeof(JvmtiTagHashmapEntry*) + duke@435: entry_count()*sizeof(JvmtiTagHashmapEntry))/K; duke@435: duke@435: int weak_globals_usage = (int)(JNIHandles::weak_global_handle_memory_usage()/K); duke@435: tty->print_cr(", %d entries (%d KB) ]", duke@435: entry_count(), hashmap_usage, weak_globals_usage); duke@435: } duke@435: duke@435: // compute threshold for the next trace message duke@435: void JvmtiTagHashmap::compute_next_trace_threshold() { duke@435: if (trace_threshold() < medium_trace_threshold) { duke@435: _trace_threshold += small_trace_threshold; duke@435: } else { duke@435: if (trace_threshold() < large_trace_threshold) { duke@435: _trace_threshold += medium_trace_threshold; duke@435: } else { duke@435: _trace_threshold += large_trace_threshold; duke@435: } duke@435: } duke@435: } duke@435: duke@435: // memory region for young generation duke@435: MemRegion JvmtiTagMap::_young_gen; duke@435: duke@435: // get the memory region used for the young generation duke@435: void JvmtiTagMap::get_young_generation() { ysr@777: CollectedHeap* ch = Universe::heap(); ysr@777: switch (ch->kind()) { ysr@777: case (CollectedHeap::GenCollectedHeap): { ysr@777: _young_gen = ((GenCollectedHeap*)ch)->get_gen(0)->reserved(); ysr@777: break; ysr@777: } duke@435: #ifndef SERIALGC ysr@777: case (CollectedHeap::ParallelScavengeHeap): { ysr@777: _young_gen = ((ParallelScavengeHeap*)ch)->young_gen()->reserved(); ysr@777: break; ysr@777: } ysr@777: case (CollectedHeap::G1CollectedHeap): { ysr@777: // Until a more satisfactory solution is implemented, all ysr@777: // oops in the tag map will require rehash at each gc. ysr@777: // This is a correct, if extremely inefficient solution. ysr@777: // See RFE 6621729 for related commentary. ysr@777: _young_gen = ch->reserved_region(); ysr@777: break; ysr@777: } ysr@777: #endif // !SERIALGC ysr@777: default: ysr@777: ShouldNotReachHere(); duke@435: } duke@435: } duke@435: duke@435: // returns true if oop is in the young generation duke@435: inline bool JvmtiTagMap::is_in_young(oop o) { duke@435: assert(_young_gen.start() != NULL, "checking"); duke@435: void* p = (void*)o; duke@435: bool in_young = _young_gen.contains(p); duke@435: return in_young; duke@435: } duke@435: duke@435: // returns the appropriate hashmap for a given object duke@435: inline JvmtiTagHashmap* JvmtiTagMap::hashmap_for(oop o) { duke@435: if (is_in_young(o)) { duke@435: return _hashmap[0]; duke@435: } else { duke@435: return _hashmap[1]; duke@435: } duke@435: } duke@435: duke@435: duke@435: // create a JvmtiTagMap duke@435: JvmtiTagMap::JvmtiTagMap(JvmtiEnv* env) : duke@435: _env(env), duke@435: _lock(Mutex::nonleaf+2, "JvmtiTagMap._lock", false), duke@435: _free_entries(NULL), duke@435: _free_entries_count(0) duke@435: { duke@435: assert(JvmtiThreadState_lock->is_locked(), "sanity check"); duke@435: assert(((JvmtiEnvBase *)env)->tag_map() == NULL, "tag map already exists for environment"); duke@435: duke@435: // create the hashmaps duke@435: for (int i=0; iset_tag_map(this); duke@435: } duke@435: duke@435: duke@435: // destroy a JvmtiTagMap duke@435: JvmtiTagMap::~JvmtiTagMap() { duke@435: duke@435: // no lock acquired as we assume the enclosing environment is duke@435: // also being destroryed. duke@435: ((JvmtiEnvBase *)_env)->set_tag_map(NULL); duke@435: duke@435: // iterate over the hashmaps and destroy each of the entries duke@435: for (int i=0; itable(); duke@435: for (int j=0; jsize(); j++) { duke@435: JvmtiTagHashmapEntry *entry = table[j]; duke@435: while (entry != NULL) { duke@435: JvmtiTagHashmapEntry* next = entry->next(); duke@435: jweak ref = entry->object(); duke@435: JNIHandles::destroy_weak_global(ref); duke@435: delete entry; duke@435: entry = next; duke@435: } duke@435: } duke@435: duke@435: // finally destroy the hashmap duke@435: delete hashmap; duke@435: } duke@435: duke@435: // remove any entries on the free list duke@435: JvmtiTagHashmapEntry* entry = _free_entries; duke@435: while (entry != NULL) { duke@435: JvmtiTagHashmapEntry* next = entry->next(); duke@435: delete entry; duke@435: entry = next; duke@435: } duke@435: } duke@435: duke@435: // create a hashmap entry duke@435: // - if there's an entry on the (per-environment) free list then this duke@435: // is returned. Otherwise an new entry is allocated. duke@435: JvmtiTagHashmapEntry* JvmtiTagMap::create_entry(jweak ref, jlong tag) { duke@435: assert(Thread::current()->is_VM_thread() || is_locked(), "checking"); duke@435: JvmtiTagHashmapEntry* entry; duke@435: if (_free_entries == NULL) { duke@435: entry = new JvmtiTagHashmapEntry(ref, tag); duke@435: } else { duke@435: assert(_free_entries_count > 0, "mismatched _free_entries_count"); duke@435: _free_entries_count--; duke@435: entry = _free_entries; duke@435: _free_entries = entry->next(); duke@435: entry->init(ref, tag); duke@435: } duke@435: return entry; duke@435: } duke@435: duke@435: // destroy an entry by returning it to the free list duke@435: void JvmtiTagMap::destroy_entry(JvmtiTagHashmapEntry* entry) { duke@435: assert(SafepointSynchronize::is_at_safepoint() || is_locked(), "checking"); duke@435: // limit the size of the free list duke@435: if (_free_entries_count >= max_free_entries) { duke@435: delete entry; duke@435: } else { duke@435: entry->set_next(_free_entries); duke@435: _free_entries = entry; duke@435: _free_entries_count++; duke@435: } duke@435: } duke@435: duke@435: // returns the tag map for the given environments. If the tag map duke@435: // doesn't exist then it is created. duke@435: JvmtiTagMap* JvmtiTagMap::tag_map_for(JvmtiEnv* env) { duke@435: JvmtiTagMap* tag_map = ((JvmtiEnvBase *)env)->tag_map(); duke@435: if (tag_map == NULL) { duke@435: MutexLocker mu(JvmtiThreadState_lock); duke@435: tag_map = ((JvmtiEnvBase *)env)->tag_map(); duke@435: if (tag_map == NULL) { duke@435: tag_map = new JvmtiTagMap(env); duke@435: } duke@435: } else { duke@435: CHECK_UNHANDLED_OOPS_ONLY(Thread::current()->clear_unhandled_oops()); duke@435: } duke@435: return tag_map; duke@435: } duke@435: duke@435: // iterate over all entries in the tag map. duke@435: void JvmtiTagMap::entry_iterate(JvmtiTagHashmapEntryClosure* closure) { duke@435: for (int i=0; ientry_iterate(closure); duke@435: } duke@435: } duke@435: duke@435: // returns true if the hashmaps are empty duke@435: bool JvmtiTagMap::is_empty() { duke@435: assert(SafepointSynchronize::is_at_safepoint() || is_locked(), "checking"); duke@435: assert(n_hashmaps == 2, "not implemented"); duke@435: return ((_hashmap[0]->entry_count() == 0) && (_hashmap[1]->entry_count() == 0)); duke@435: } duke@435: duke@435: duke@435: // Return the tag value for an object, or 0 if the object is duke@435: // not tagged duke@435: // duke@435: static inline jlong tag_for(JvmtiTagMap* tag_map, oop o) { duke@435: JvmtiTagHashmapEntry* entry = tag_map->hashmap_for(o)->find(o); duke@435: if (entry == NULL) { duke@435: return 0; duke@435: } else { duke@435: return entry->tag(); duke@435: } duke@435: } duke@435: duke@435: // If the object is a java.lang.Class then return the klassOop, duke@435: // otherwise return the original object duke@435: static inline oop klassOop_if_java_lang_Class(oop o) { duke@435: if (o->klass() == SystemDictionary::class_klass()) { duke@435: if (!java_lang_Class::is_primitive(o)) { duke@435: o = (oop)java_lang_Class::as_klassOop(o); duke@435: assert(o != NULL, "class for non-primitive mirror must exist"); duke@435: } duke@435: } duke@435: return o; duke@435: } duke@435: duke@435: // A CallbackWrapper is a support class for querying and tagging an object duke@435: // around a callback to a profiler. The constructor does pre-callback duke@435: // work to get the tag value, klass tag value, ... and the destructor duke@435: // does the post-callback work of tagging or untagging the object. duke@435: // duke@435: // { duke@435: // CallbackWrapper wrapper(tag_map, o); duke@435: // duke@435: // (*callback)(wrapper.klass_tag(), wrapper.obj_size(), wrapper.obj_tag_p(), ...) duke@435: // duke@435: // } // wrapper goes out of scope here which results in the destructor duke@435: // checking to see if the object has been tagged, untagged, or the duke@435: // tag value has changed. duke@435: // duke@435: class CallbackWrapper : public StackObj { duke@435: private: duke@435: JvmtiTagMap* _tag_map; duke@435: JvmtiTagHashmap* _hashmap; duke@435: JvmtiTagHashmapEntry* _entry; duke@435: oop _o; duke@435: jlong _obj_size; duke@435: jlong _obj_tag; duke@435: klassOop _klass; // the object's class duke@435: jlong _klass_tag; duke@435: duke@435: protected: duke@435: JvmtiTagMap* tag_map() const { return _tag_map; } duke@435: duke@435: // invoked post-callback to tag, untag, or update the tag of an object duke@435: void inline post_callback_tag_update(oop o, JvmtiTagHashmap* hashmap, duke@435: JvmtiTagHashmapEntry* entry, jlong obj_tag); duke@435: public: duke@435: CallbackWrapper(JvmtiTagMap* tag_map, oop o) { duke@435: assert(Thread::current()->is_VM_thread() || tag_map->is_locked(), duke@435: "MT unsafe or must be VM thread"); duke@435: duke@435: // for Classes the klassOop is tagged duke@435: _o = klassOop_if_java_lang_Class(o); duke@435: duke@435: // object size duke@435: _obj_size = _o->size() * wordSize; duke@435: duke@435: // record the context duke@435: _tag_map = tag_map; duke@435: _hashmap = tag_map->hashmap_for(_o); duke@435: _entry = _hashmap->find(_o); duke@435: duke@435: // get object tag duke@435: _obj_tag = (_entry == NULL) ? 0 : _entry->tag(); duke@435: duke@435: // get the class and the class's tag value duke@435: if (_o == o) { duke@435: _klass = _o->klass(); duke@435: } else { duke@435: // if the object represents a runtime class then use the duke@435: // tag for java.lang.Class duke@435: _klass = SystemDictionary::class_klass(); duke@435: } duke@435: _klass_tag = tag_for(tag_map, _klass); duke@435: } duke@435: duke@435: ~CallbackWrapper() { duke@435: post_callback_tag_update(_o, _hashmap, _entry, _obj_tag); duke@435: } duke@435: duke@435: inline jlong* obj_tag_p() { return &_obj_tag; } duke@435: inline jlong obj_size() const { return _obj_size; } duke@435: inline jlong obj_tag() const { return _obj_tag; } duke@435: inline klassOop klass() const { return _klass; } duke@435: inline jlong klass_tag() const { return _klass_tag; } duke@435: }; duke@435: duke@435: duke@435: duke@435: // callback post-callback to tag, untag, or update the tag of an object duke@435: void inline CallbackWrapper::post_callback_tag_update(oop o, duke@435: JvmtiTagHashmap* hashmap, duke@435: JvmtiTagHashmapEntry* entry, duke@435: jlong obj_tag) { duke@435: if (entry == NULL) { duke@435: if (obj_tag != 0) { duke@435: // callback has tagged the object duke@435: assert(Thread::current()->is_VM_thread(), "must be VMThread"); duke@435: HandleMark hm; duke@435: Handle h(o); duke@435: jweak ref = JNIHandles::make_weak_global(h); duke@435: entry = tag_map()->create_entry(ref, obj_tag); duke@435: hashmap->add(o, entry); duke@435: } duke@435: } else { duke@435: // object was previously tagged - the callback may have untagged duke@435: // the object or changed the tag value duke@435: if (obj_tag == 0) { duke@435: jweak ref = entry->object(); duke@435: duke@435: JvmtiTagHashmapEntry* entry_removed = hashmap->remove(o); duke@435: assert(entry_removed == entry, "checking"); duke@435: tag_map()->destroy_entry(entry); duke@435: duke@435: JNIHandles::destroy_weak_global(ref); duke@435: } else { duke@435: if (obj_tag != entry->tag()) { duke@435: entry->set_tag(obj_tag); duke@435: } duke@435: } duke@435: } duke@435: } duke@435: duke@435: // An extended CallbackWrapper used when reporting an object reference duke@435: // to the agent. duke@435: // duke@435: // { duke@435: // TwoOopCallbackWrapper wrapper(tag_map, referrer, o); duke@435: // duke@435: // (*callback)(wrapper.klass_tag(), duke@435: // wrapper.obj_size(), duke@435: // wrapper.obj_tag_p() duke@435: // wrapper.referrer_tag_p(), ...) duke@435: // duke@435: // } // wrapper goes out of scope here which results in the destructor duke@435: // checking to see if the referrer object has been tagged, untagged, duke@435: // or the tag value has changed. duke@435: // duke@435: class TwoOopCallbackWrapper : public CallbackWrapper { duke@435: private: duke@435: bool _is_reference_to_self; duke@435: JvmtiTagHashmap* _referrer_hashmap; duke@435: JvmtiTagHashmapEntry* _referrer_entry; duke@435: oop _referrer; duke@435: jlong _referrer_obj_tag; duke@435: jlong _referrer_klass_tag; duke@435: jlong* _referrer_tag_p; duke@435: duke@435: bool is_reference_to_self() const { return _is_reference_to_self; } duke@435: duke@435: public: duke@435: TwoOopCallbackWrapper(JvmtiTagMap* tag_map, oop referrer, oop o) : duke@435: CallbackWrapper(tag_map, o) duke@435: { duke@435: // self reference needs to be handled in a special way duke@435: _is_reference_to_self = (referrer == o); duke@435: duke@435: if (_is_reference_to_self) { duke@435: _referrer_klass_tag = klass_tag(); duke@435: _referrer_tag_p = obj_tag_p(); duke@435: } else { duke@435: // for Classes the klassOop is tagged duke@435: _referrer = klassOop_if_java_lang_Class(referrer); duke@435: // record the context duke@435: _referrer_hashmap = tag_map->hashmap_for(_referrer); duke@435: _referrer_entry = _referrer_hashmap->find(_referrer); duke@435: duke@435: // get object tag duke@435: _referrer_obj_tag = (_referrer_entry == NULL) ? 0 : _referrer_entry->tag(); duke@435: _referrer_tag_p = &_referrer_obj_tag; duke@435: duke@435: // get referrer class tag. duke@435: klassOop k = (_referrer == referrer) ? // Check if referrer is a class... duke@435: _referrer->klass() // No, just get its class duke@435: : SystemDictionary::class_klass(); // Yes, its class is Class duke@435: _referrer_klass_tag = tag_for(tag_map, k); duke@435: } duke@435: } duke@435: duke@435: ~TwoOopCallbackWrapper() { duke@435: if (!is_reference_to_self()){ duke@435: post_callback_tag_update(_referrer, duke@435: _referrer_hashmap, duke@435: _referrer_entry, duke@435: _referrer_obj_tag); duke@435: } duke@435: } duke@435: duke@435: // address of referrer tag duke@435: // (for a self reference this will return the same thing as obj_tag_p()) duke@435: inline jlong* referrer_tag_p() { return _referrer_tag_p; } duke@435: duke@435: // referrer's class tag duke@435: inline jlong referrer_klass_tag() { return _referrer_klass_tag; } duke@435: }; duke@435: duke@435: // tag an object duke@435: // duke@435: // This function is performance critical. If many threads attempt to tag objects duke@435: // around the same time then it's possible that the Mutex associated with the duke@435: // tag map will be a hot lock. Eliminating this lock will not eliminate the issue duke@435: // because creating a JNI weak reference requires acquiring a global lock also. duke@435: void JvmtiTagMap::set_tag(jobject object, jlong tag) { duke@435: MutexLocker ml(lock()); duke@435: duke@435: // resolve the object duke@435: oop o = JNIHandles::resolve_non_null(object); duke@435: duke@435: // for Classes we tag the klassOop duke@435: o = klassOop_if_java_lang_Class(o); duke@435: duke@435: // see if the object is already tagged duke@435: JvmtiTagHashmap* hashmap = hashmap_for(o); duke@435: JvmtiTagHashmapEntry* entry = hashmap->find(o); duke@435: duke@435: // if the object is not already tagged then we tag it duke@435: if (entry == NULL) { duke@435: if (tag != 0) { duke@435: HandleMark hm; duke@435: Handle h(o); duke@435: jweak ref = JNIHandles::make_weak_global(h); duke@435: duke@435: // the object may have moved because make_weak_global may duke@435: // have blocked - thus it is necessary resolve the handle duke@435: // and re-hash the object. duke@435: o = h(); duke@435: entry = create_entry(ref, tag); duke@435: hashmap_for(o)->add(o, entry); duke@435: } else { duke@435: // no-op duke@435: } duke@435: } else { duke@435: // if the object is already tagged then we either update duke@435: // the tag (if a new tag value has been provided) duke@435: // or remove the object if the new tag value is 0. duke@435: // Removing the object requires that we also delete the JNI duke@435: // weak ref to the object. duke@435: if (tag == 0) { duke@435: jweak ref = entry->object(); duke@435: hashmap->remove(o); duke@435: destroy_entry(entry); duke@435: JNIHandles::destroy_weak_global(ref); duke@435: } else { duke@435: entry->set_tag(tag); duke@435: } duke@435: } duke@435: } duke@435: duke@435: // get the tag for an object duke@435: jlong JvmtiTagMap::get_tag(jobject object) { duke@435: MutexLocker ml(lock()); duke@435: duke@435: // resolve the object duke@435: oop o = JNIHandles::resolve_non_null(object); duke@435: duke@435: // for Classes get the tag from the klassOop duke@435: return tag_for(this, klassOop_if_java_lang_Class(o)); duke@435: } duke@435: duke@435: duke@435: // Helper class used to describe the static or instance fields of a class. duke@435: // For each field it holds the field index (as defined by the JVMTI specification), duke@435: // the field type, and the offset. duke@435: duke@435: class ClassFieldDescriptor: public CHeapObj { duke@435: private: duke@435: int _field_index; duke@435: int _field_offset; duke@435: char _field_type; duke@435: public: duke@435: ClassFieldDescriptor(int index, char type, int offset) : duke@435: _field_index(index), _field_type(type), _field_offset(offset) { duke@435: } duke@435: int field_index() const { return _field_index; } duke@435: char field_type() const { return _field_type; } duke@435: int field_offset() const { return _field_offset; } duke@435: }; duke@435: duke@435: class ClassFieldMap: public CHeapObj { duke@435: private: duke@435: enum { duke@435: initial_field_count = 5 duke@435: }; duke@435: duke@435: // list of field descriptors duke@435: GrowableArray* _fields; duke@435: duke@435: // constructor duke@435: ClassFieldMap(); duke@435: duke@435: // add a field duke@435: void add(int index, char type, int offset); duke@435: duke@435: // returns the field count for the given class duke@435: static int compute_field_count(instanceKlassHandle ikh); duke@435: duke@435: public: duke@435: ~ClassFieldMap(); duke@435: duke@435: // access duke@435: int field_count() { return _fields->length(); } duke@435: ClassFieldDescriptor* field_at(int i) { return _fields->at(i); } duke@435: duke@435: // functions to create maps of static or instance fields duke@435: static ClassFieldMap* create_map_of_static_fields(klassOop k); duke@435: static ClassFieldMap* create_map_of_instance_fields(oop obj); duke@435: }; duke@435: duke@435: ClassFieldMap::ClassFieldMap() { duke@435: _fields = new (ResourceObj::C_HEAP) GrowableArray(initial_field_count, true); duke@435: } duke@435: duke@435: ClassFieldMap::~ClassFieldMap() { duke@435: for (int i=0; i<_fields->length(); i++) { duke@435: delete _fields->at(i); duke@435: } duke@435: delete _fields; duke@435: } duke@435: duke@435: void ClassFieldMap::add(int index, char type, int offset) { duke@435: ClassFieldDescriptor* field = new ClassFieldDescriptor(index, type, offset); duke@435: _fields->append(field); duke@435: } duke@435: duke@435: // Returns a heap allocated ClassFieldMap to describe the static fields duke@435: // of the given class. duke@435: // duke@435: ClassFieldMap* ClassFieldMap::create_map_of_static_fields(klassOop k) { duke@435: HandleMark hm; duke@435: instanceKlassHandle ikh = instanceKlassHandle(Thread::current(), k); duke@435: duke@435: // create the field map duke@435: ClassFieldMap* field_map = new ClassFieldMap(); duke@435: duke@435: FilteredFieldStream f(ikh, false, false); duke@435: int max_field_index = f.field_count()-1; duke@435: duke@435: int index = 0; duke@435: for (FilteredFieldStream fld(ikh, true, true); !fld.eos(); fld.next(), index++) { duke@435: // ignore instance fields duke@435: if (!fld.access_flags().is_static()) { duke@435: continue; duke@435: } duke@435: field_map->add(max_field_index - index, fld.signature()->byte_at(0), fld.offset()); duke@435: } duke@435: return field_map; duke@435: } duke@435: duke@435: // Returns a heap allocated ClassFieldMap to describe the instance fields duke@435: // of the given class. All instance fields are included (this means public duke@435: // and private fields declared in superclasses and superinterfaces too). duke@435: // duke@435: ClassFieldMap* ClassFieldMap::create_map_of_instance_fields(oop obj) { duke@435: HandleMark hm; duke@435: instanceKlassHandle ikh = instanceKlassHandle(Thread::current(), obj->klass()); duke@435: duke@435: // create the field map duke@435: ClassFieldMap* field_map = new ClassFieldMap(); duke@435: duke@435: FilteredFieldStream f(ikh, false, false); duke@435: duke@435: int max_field_index = f.field_count()-1; duke@435: duke@435: int index = 0; duke@435: for (FilteredFieldStream fld(ikh, false, false); !fld.eos(); fld.next(), index++) { duke@435: // ignore static fields duke@435: if (fld.access_flags().is_static()) { duke@435: continue; duke@435: } duke@435: field_map->add(max_field_index - index, fld.signature()->byte_at(0), fld.offset()); duke@435: } duke@435: duke@435: return field_map; duke@435: } duke@435: duke@435: // Helper class used to cache a ClassFileMap for the instance fields of duke@435: // a cache. A JvmtiCachedClassFieldMap can be cached by an instanceKlass during duke@435: // heap iteration and avoid creating a field map for each object in the heap duke@435: // (only need to create the map when the first instance of a class is encountered). duke@435: // duke@435: class JvmtiCachedClassFieldMap : public CHeapObj { duke@435: private: duke@435: enum { duke@435: initial_class_count = 200 duke@435: }; duke@435: ClassFieldMap* _field_map; duke@435: duke@435: ClassFieldMap* field_map() const { return _field_map; } duke@435: duke@435: JvmtiCachedClassFieldMap(ClassFieldMap* field_map); duke@435: ~JvmtiCachedClassFieldMap(); duke@435: duke@435: static GrowableArray* _class_list; duke@435: static void add_to_class_list(instanceKlass* ik); duke@435: duke@435: public: duke@435: // returns the field map for a given object (returning map cached duke@435: // by instanceKlass if possible duke@435: static ClassFieldMap* get_map_of_instance_fields(oop obj); duke@435: duke@435: // removes the field map from all instanceKlasses - should be duke@435: // called before VM operation completes duke@435: static void clear_cache(); duke@435: duke@435: // returns the number of ClassFieldMap cached by instanceKlasses duke@435: static int cached_field_map_count(); duke@435: }; duke@435: duke@435: GrowableArray* JvmtiCachedClassFieldMap::_class_list; duke@435: duke@435: JvmtiCachedClassFieldMap::JvmtiCachedClassFieldMap(ClassFieldMap* field_map) { duke@435: _field_map = field_map; duke@435: } duke@435: duke@435: JvmtiCachedClassFieldMap::~JvmtiCachedClassFieldMap() { duke@435: if (_field_map != NULL) { duke@435: delete _field_map; duke@435: } duke@435: } duke@435: duke@435: // Marker class to ensure that the class file map cache is only used in a defined duke@435: // scope. duke@435: class ClassFieldMapCacheMark : public StackObj { duke@435: private: duke@435: static bool _is_active; duke@435: public: duke@435: ClassFieldMapCacheMark() { duke@435: assert(Thread::current()->is_VM_thread(), "must be VMThread"); duke@435: assert(JvmtiCachedClassFieldMap::cached_field_map_count() == 0, "cache not empty"); duke@435: assert(!_is_active, "ClassFieldMapCacheMark cannot be nested"); duke@435: _is_active = true; duke@435: } duke@435: ~ClassFieldMapCacheMark() { duke@435: JvmtiCachedClassFieldMap::clear_cache(); duke@435: _is_active = false; duke@435: } duke@435: static bool is_active() { return _is_active; } duke@435: }; duke@435: duke@435: bool ClassFieldMapCacheMark::_is_active; duke@435: duke@435: duke@435: // record that the given instanceKlass is caching a field map duke@435: void JvmtiCachedClassFieldMap::add_to_class_list(instanceKlass* ik) { duke@435: if (_class_list == NULL) { duke@435: _class_list = new (ResourceObj::C_HEAP) GrowableArray(initial_class_count, true); duke@435: } duke@435: _class_list->push(ik); duke@435: } duke@435: duke@435: // returns the instance field map for the given object duke@435: // (returns field map cached by the instanceKlass if possible) duke@435: ClassFieldMap* JvmtiCachedClassFieldMap::get_map_of_instance_fields(oop obj) { duke@435: assert(Thread::current()->is_VM_thread(), "must be VMThread"); duke@435: assert(ClassFieldMapCacheMark::is_active(), "ClassFieldMapCacheMark not active"); duke@435: duke@435: klassOop k = obj->klass(); duke@435: instanceKlass* ik = instanceKlass::cast(k); duke@435: duke@435: // return cached map if possible duke@435: JvmtiCachedClassFieldMap* cached_map = ik->jvmti_cached_class_field_map(); duke@435: if (cached_map != NULL) { duke@435: assert(cached_map->field_map() != NULL, "missing field list"); duke@435: return cached_map->field_map(); duke@435: } else { duke@435: ClassFieldMap* field_map = ClassFieldMap::create_map_of_instance_fields(obj); duke@435: cached_map = new JvmtiCachedClassFieldMap(field_map); duke@435: ik->set_jvmti_cached_class_field_map(cached_map); duke@435: add_to_class_list(ik); duke@435: return field_map; duke@435: } duke@435: } duke@435: duke@435: // remove the fields maps cached from all instanceKlasses duke@435: void JvmtiCachedClassFieldMap::clear_cache() { duke@435: assert(Thread::current()->is_VM_thread(), "must be VMThread"); duke@435: if (_class_list != NULL) { duke@435: for (int i = 0; i < _class_list->length(); i++) { duke@435: instanceKlass* ik = _class_list->at(i); duke@435: JvmtiCachedClassFieldMap* cached_map = ik->jvmti_cached_class_field_map(); duke@435: assert(cached_map != NULL, "should not be NULL"); duke@435: ik->set_jvmti_cached_class_field_map(NULL); duke@435: delete cached_map; // deletes the encapsulated field map duke@435: } duke@435: delete _class_list; duke@435: _class_list = NULL; duke@435: } duke@435: } duke@435: duke@435: // returns the number of ClassFieldMap cached by instanceKlasses duke@435: int JvmtiCachedClassFieldMap::cached_field_map_count() { duke@435: return (_class_list == NULL) ? 0 : _class_list->length(); duke@435: } duke@435: duke@435: // helper function to indicate if an object is filtered by its tag or class tag duke@435: static inline bool is_filtered_by_heap_filter(jlong obj_tag, duke@435: jlong klass_tag, duke@435: int heap_filter) { duke@435: // apply the heap filter duke@435: if (obj_tag != 0) { duke@435: // filter out tagged objects duke@435: if (heap_filter & JVMTI_HEAP_FILTER_TAGGED) return true; duke@435: } else { duke@435: // filter out untagged objects duke@435: if (heap_filter & JVMTI_HEAP_FILTER_UNTAGGED) return true; duke@435: } duke@435: if (klass_tag != 0) { duke@435: // filter out objects with tagged classes duke@435: if (heap_filter & JVMTI_HEAP_FILTER_CLASS_TAGGED) return true; duke@435: } else { duke@435: // filter out objects with untagged classes. duke@435: if (heap_filter & JVMTI_HEAP_FILTER_CLASS_UNTAGGED) return true; duke@435: } duke@435: return false; duke@435: } duke@435: duke@435: // helper function to indicate if an object is filtered by a klass filter duke@435: static inline bool is_filtered_by_klass_filter(oop obj, KlassHandle klass_filter) { duke@435: if (!klass_filter.is_null()) { duke@435: if (obj->klass() != klass_filter()) { duke@435: return true; duke@435: } duke@435: } duke@435: return false; duke@435: } duke@435: duke@435: // helper function to tell if a field is a primitive field or not duke@435: static inline bool is_primitive_field_type(char type) { duke@435: return (type != 'L' && type != '['); duke@435: } duke@435: duke@435: // helper function to copy the value from location addr to jvalue. duke@435: static inline void copy_to_jvalue(jvalue *v, address addr, jvmtiPrimitiveType value_type) { duke@435: switch (value_type) { duke@435: case JVMTI_PRIMITIVE_TYPE_BOOLEAN : { v->z = *(jboolean*)addr; break; } duke@435: case JVMTI_PRIMITIVE_TYPE_BYTE : { v->b = *(jbyte*)addr; break; } duke@435: case JVMTI_PRIMITIVE_TYPE_CHAR : { v->c = *(jchar*)addr; break; } duke@435: case JVMTI_PRIMITIVE_TYPE_SHORT : { v->s = *(jshort*)addr; break; } duke@435: case JVMTI_PRIMITIVE_TYPE_INT : { v->i = *(jint*)addr; break; } duke@435: case JVMTI_PRIMITIVE_TYPE_LONG : { v->j = *(jlong*)addr; break; } duke@435: case JVMTI_PRIMITIVE_TYPE_FLOAT : { v->f = *(jfloat*)addr; break; } duke@435: case JVMTI_PRIMITIVE_TYPE_DOUBLE : { v->d = *(jdouble*)addr; break; } duke@435: default: ShouldNotReachHere(); duke@435: } duke@435: } duke@435: duke@435: // helper function to invoke string primitive value callback duke@435: // returns visit control flags duke@435: static jint invoke_string_value_callback(jvmtiStringPrimitiveValueCallback cb, duke@435: CallbackWrapper* wrapper, duke@435: oop str, duke@435: void* user_data) duke@435: { duke@435: assert(str->klass() == SystemDictionary::string_klass(), "not a string"); duke@435: duke@435: // get the string value and length duke@435: // (string value may be offset from the base) duke@435: int s_len = java_lang_String::length(str); duke@435: typeArrayOop s_value = java_lang_String::value(str); duke@435: int s_offset = java_lang_String::offset(str); duke@435: jchar* value; duke@435: if (s_len > 0) { duke@435: value = s_value->char_at_addr(s_offset); duke@435: } else { duke@435: value = (jchar*) s_value->base(T_CHAR); duke@435: } duke@435: duke@435: // invoke the callback duke@435: return (*cb)(wrapper->klass_tag(), duke@435: wrapper->obj_size(), duke@435: wrapper->obj_tag_p(), duke@435: value, duke@435: (jint)s_len, duke@435: user_data); duke@435: } duke@435: duke@435: // helper function to invoke string primitive value callback duke@435: // returns visit control flags duke@435: static jint invoke_array_primitive_value_callback(jvmtiArrayPrimitiveValueCallback cb, duke@435: CallbackWrapper* wrapper, duke@435: oop obj, duke@435: void* user_data) duke@435: { duke@435: assert(obj->is_typeArray(), "not a primitive array"); duke@435: duke@435: // get base address of first element duke@435: typeArrayOop array = typeArrayOop(obj); duke@435: BasicType type = typeArrayKlass::cast(array->klass())->element_type(); duke@435: void* elements = array->base(type); duke@435: duke@435: // jvmtiPrimitiveType is defined so this mapping is always correct duke@435: jvmtiPrimitiveType elem_type = (jvmtiPrimitiveType)type2char(type); duke@435: duke@435: return (*cb)(wrapper->klass_tag(), duke@435: wrapper->obj_size(), duke@435: wrapper->obj_tag_p(), duke@435: (jint)array->length(), duke@435: elem_type, duke@435: elements, duke@435: user_data); duke@435: } duke@435: duke@435: // helper function to invoke the primitive field callback for all static fields duke@435: // of a given class duke@435: static jint invoke_primitive_field_callback_for_static_fields duke@435: (CallbackWrapper* wrapper, duke@435: oop obj, duke@435: jvmtiPrimitiveFieldCallback cb, duke@435: void* user_data) duke@435: { duke@435: // for static fields only the index will be set duke@435: static jvmtiHeapReferenceInfo reference_info = { 0 }; duke@435: duke@435: assert(obj->klass() == SystemDictionary::class_klass(), "not a class"); duke@435: if (java_lang_Class::is_primitive(obj)) { duke@435: return 0; duke@435: } duke@435: klassOop k = java_lang_Class::as_klassOop(obj); duke@435: Klass* klass = k->klass_part(); duke@435: duke@435: // ignore classes for object and type arrays duke@435: if (!klass->oop_is_instance()) { duke@435: return 0; duke@435: } duke@435: duke@435: // ignore classes which aren't linked yet duke@435: instanceKlass* ik = instanceKlass::cast(k); duke@435: if (!ik->is_linked()) { duke@435: return 0; duke@435: } duke@435: duke@435: // get the field map duke@435: ClassFieldMap* field_map = ClassFieldMap::create_map_of_static_fields(k); duke@435: duke@435: // invoke the callback for each static primitive field duke@435: for (int i=0; ifield_count(); i++) { duke@435: ClassFieldDescriptor* field = field_map->field_at(i); duke@435: duke@435: // ignore non-primitive fields duke@435: char type = field->field_type(); duke@435: if (!is_primitive_field_type(type)) { duke@435: continue; duke@435: } duke@435: // one-to-one mapping duke@435: jvmtiPrimitiveType value_type = (jvmtiPrimitiveType)type; duke@435: duke@435: // get offset and field value duke@435: int offset = field->field_offset(); duke@435: address addr = (address)k + offset; duke@435: jvalue value; duke@435: copy_to_jvalue(&value, addr, value_type); duke@435: duke@435: // field index duke@435: reference_info.field.index = field->field_index(); duke@435: duke@435: // invoke the callback duke@435: jint res = (*cb)(JVMTI_HEAP_REFERENCE_STATIC_FIELD, duke@435: &reference_info, duke@435: wrapper->klass_tag(), duke@435: wrapper->obj_tag_p(), duke@435: value, duke@435: value_type, duke@435: user_data); duke@435: if (res & JVMTI_VISIT_ABORT) { duke@435: delete field_map; duke@435: return res; duke@435: } duke@435: } duke@435: duke@435: delete field_map; duke@435: return 0; duke@435: } duke@435: duke@435: // helper function to invoke the primitive field callback for all instance fields duke@435: // of a given object duke@435: static jint invoke_primitive_field_callback_for_instance_fields( duke@435: CallbackWrapper* wrapper, duke@435: oop obj, duke@435: jvmtiPrimitiveFieldCallback cb, duke@435: void* user_data) duke@435: { duke@435: // for instance fields only the index will be set duke@435: static jvmtiHeapReferenceInfo reference_info = { 0 }; duke@435: duke@435: // get the map of the instance fields duke@435: ClassFieldMap* fields = JvmtiCachedClassFieldMap::get_map_of_instance_fields(obj); duke@435: duke@435: // invoke the callback for each instance primitive field duke@435: for (int i=0; ifield_count(); i++) { duke@435: ClassFieldDescriptor* field = fields->field_at(i); duke@435: duke@435: // ignore non-primitive fields duke@435: char type = field->field_type(); duke@435: if (!is_primitive_field_type(type)) { duke@435: continue; duke@435: } duke@435: // one-to-one mapping duke@435: jvmtiPrimitiveType value_type = (jvmtiPrimitiveType)type; duke@435: duke@435: // get offset and field value duke@435: int offset = field->field_offset(); duke@435: address addr = (address)obj + offset; duke@435: jvalue value; duke@435: copy_to_jvalue(&value, addr, value_type); duke@435: duke@435: // field index duke@435: reference_info.field.index = field->field_index(); duke@435: duke@435: // invoke the callback duke@435: jint res = (*cb)(JVMTI_HEAP_REFERENCE_FIELD, duke@435: &reference_info, duke@435: wrapper->klass_tag(), duke@435: wrapper->obj_tag_p(), duke@435: value, duke@435: value_type, duke@435: user_data); duke@435: if (res & JVMTI_VISIT_ABORT) { duke@435: return res; duke@435: } duke@435: } duke@435: return 0; duke@435: } duke@435: duke@435: duke@435: // VM operation to iterate over all objects in the heap (both reachable duke@435: // and unreachable) duke@435: class VM_HeapIterateOperation: public VM_Operation { duke@435: private: duke@435: ObjectClosure* _blk; duke@435: public: duke@435: VM_HeapIterateOperation(ObjectClosure* blk) { _blk = blk; } duke@435: duke@435: VMOp_Type type() const { return VMOp_HeapIterateOperation; } duke@435: void doit() { duke@435: // allows class files maps to be cached during iteration duke@435: ClassFieldMapCacheMark cm; duke@435: duke@435: // make sure that heap is parsable (fills TLABs with filler objects) duke@435: Universe::heap()->ensure_parsability(false); // no need to retire TLABs duke@435: duke@435: // Verify heap before iteration - if the heap gets corrupted then duke@435: // JVMTI's IterateOverHeap will crash. duke@435: if (VerifyBeforeIteration) { duke@435: Universe::verify(); duke@435: } duke@435: duke@435: // do the iteration duke@435: Universe::heap()->object_iterate(_blk); duke@435: duke@435: // when sharing is enabled we must iterate over the shared spaces duke@435: if (UseSharedSpaces) { duke@435: GenCollectedHeap* gch = GenCollectedHeap::heap(); duke@435: CompactingPermGenGen* gen = (CompactingPermGenGen*)gch->perm_gen(); duke@435: gen->ro_space()->object_iterate(_blk); duke@435: gen->rw_space()->object_iterate(_blk); duke@435: } duke@435: } duke@435: duke@435: }; duke@435: duke@435: duke@435: // An ObjectClosure used to support the deprecated IterateOverHeap and duke@435: // IterateOverInstancesOfClass functions duke@435: class IterateOverHeapObjectClosure: public ObjectClosure { duke@435: private: duke@435: JvmtiTagMap* _tag_map; duke@435: KlassHandle _klass; duke@435: jvmtiHeapObjectFilter _object_filter; duke@435: jvmtiHeapObjectCallback _heap_object_callback; duke@435: const void* _user_data; duke@435: duke@435: // accessors duke@435: JvmtiTagMap* tag_map() const { return _tag_map; } duke@435: jvmtiHeapObjectFilter object_filter() const { return _object_filter; } duke@435: jvmtiHeapObjectCallback object_callback() const { return _heap_object_callback; } duke@435: KlassHandle klass() const { return _klass; } duke@435: const void* user_data() const { return _user_data; } duke@435: duke@435: // indicates if iteration has been aborted duke@435: bool _iteration_aborted; duke@435: bool is_iteration_aborted() const { return _iteration_aborted; } duke@435: void set_iteration_aborted(bool aborted) { _iteration_aborted = aborted; } duke@435: duke@435: public: duke@435: IterateOverHeapObjectClosure(JvmtiTagMap* tag_map, duke@435: KlassHandle klass, duke@435: jvmtiHeapObjectFilter object_filter, duke@435: jvmtiHeapObjectCallback heap_object_callback, duke@435: const void* user_data) : duke@435: _tag_map(tag_map), duke@435: _klass(klass), duke@435: _object_filter(object_filter), duke@435: _heap_object_callback(heap_object_callback), duke@435: _user_data(user_data), duke@435: _iteration_aborted(false) duke@435: { duke@435: } duke@435: duke@435: void do_object(oop o); duke@435: }; duke@435: duke@435: // invoked for each object in the heap duke@435: void IterateOverHeapObjectClosure::do_object(oop o) { duke@435: // check if iteration has been halted duke@435: if (is_iteration_aborted()) return; duke@435: duke@435: // ignore any objects that aren't visible to profiler duke@435: if (!ServiceUtil::visible_oop(o)) return; duke@435: duke@435: // instanceof check when filtering by klass duke@435: if (!klass().is_null() && !o->is_a(klass()())) { duke@435: return; duke@435: } duke@435: // prepare for the calllback duke@435: CallbackWrapper wrapper(tag_map(), o); duke@435: duke@435: // if the object is tagged and we're only interested in untagged objects duke@435: // then don't invoke the callback. Similiarly, if the object is untagged duke@435: // and we're only interested in tagged objects we skip the callback. duke@435: if (wrapper.obj_tag() != 0) { duke@435: if (object_filter() == JVMTI_HEAP_OBJECT_UNTAGGED) return; duke@435: } else { duke@435: if (object_filter() == JVMTI_HEAP_OBJECT_TAGGED) return; duke@435: } duke@435: duke@435: // invoke the agent's callback duke@435: jvmtiIterationControl control = (*object_callback())(wrapper.klass_tag(), duke@435: wrapper.obj_size(), duke@435: wrapper.obj_tag_p(), duke@435: (void*)user_data()); duke@435: if (control == JVMTI_ITERATION_ABORT) { duke@435: set_iteration_aborted(true); duke@435: } duke@435: } duke@435: duke@435: // An ObjectClosure used to support the IterateThroughHeap function duke@435: class IterateThroughHeapObjectClosure: public ObjectClosure { duke@435: private: duke@435: JvmtiTagMap* _tag_map; duke@435: KlassHandle _klass; duke@435: int _heap_filter; duke@435: const jvmtiHeapCallbacks* _callbacks; duke@435: const void* _user_data; duke@435: duke@435: // accessor functions duke@435: JvmtiTagMap* tag_map() const { return _tag_map; } duke@435: int heap_filter() const { return _heap_filter; } duke@435: const jvmtiHeapCallbacks* callbacks() const { return _callbacks; } duke@435: KlassHandle klass() const { return _klass; } duke@435: const void* user_data() const { return _user_data; } duke@435: duke@435: // indicates if the iteration has been aborted duke@435: bool _iteration_aborted; duke@435: bool is_iteration_aborted() const { return _iteration_aborted; } duke@435: duke@435: // used to check the visit control flags. If the abort flag is set duke@435: // then we set the iteration aborted flag so that the iteration completes duke@435: // without processing any further objects duke@435: bool check_flags_for_abort(jint flags) { duke@435: bool is_abort = (flags & JVMTI_VISIT_ABORT) != 0; duke@435: if (is_abort) { duke@435: _iteration_aborted = true; duke@435: } duke@435: return is_abort; duke@435: } duke@435: duke@435: public: duke@435: IterateThroughHeapObjectClosure(JvmtiTagMap* tag_map, duke@435: KlassHandle klass, duke@435: int heap_filter, duke@435: const jvmtiHeapCallbacks* heap_callbacks, duke@435: const void* user_data) : duke@435: _tag_map(tag_map), duke@435: _klass(klass), duke@435: _heap_filter(heap_filter), duke@435: _callbacks(heap_callbacks), duke@435: _user_data(user_data), duke@435: _iteration_aborted(false) duke@435: { duke@435: } duke@435: duke@435: void do_object(oop o); duke@435: }; duke@435: duke@435: // invoked for each object in the heap duke@435: void IterateThroughHeapObjectClosure::do_object(oop obj) { duke@435: // check if iteration has been halted duke@435: if (is_iteration_aborted()) return; duke@435: duke@435: // ignore any objects that aren't visible to profiler duke@435: if (!ServiceUtil::visible_oop(obj)) return; duke@435: duke@435: // apply class filter duke@435: if (is_filtered_by_klass_filter(obj, klass())) return; duke@435: duke@435: // prepare for callback duke@435: CallbackWrapper wrapper(tag_map(), obj); duke@435: duke@435: // check if filtered by the heap filter duke@435: if (is_filtered_by_heap_filter(wrapper.obj_tag(), wrapper.klass_tag(), heap_filter())) { duke@435: return; duke@435: } duke@435: duke@435: // for arrays we need the length, otherwise -1 duke@435: bool is_array = obj->is_array(); duke@435: int len = is_array ? arrayOop(obj)->length() : -1; duke@435: duke@435: // invoke the object callback (if callback is provided) duke@435: if (callbacks()->heap_iteration_callback != NULL) { duke@435: jvmtiHeapIterationCallback cb = callbacks()->heap_iteration_callback; duke@435: jint res = (*cb)(wrapper.klass_tag(), duke@435: wrapper.obj_size(), duke@435: wrapper.obj_tag_p(), duke@435: (jint)len, duke@435: (void*)user_data()); duke@435: if (check_flags_for_abort(res)) return; duke@435: } duke@435: duke@435: // for objects and classes we report primitive fields if callback provided duke@435: if (callbacks()->primitive_field_callback != NULL && obj->is_instance()) { duke@435: jint res; duke@435: jvmtiPrimitiveFieldCallback cb = callbacks()->primitive_field_callback; duke@435: if (obj->klass() == SystemDictionary::class_klass()) { duke@435: res = invoke_primitive_field_callback_for_static_fields(&wrapper, duke@435: obj, duke@435: cb, duke@435: (void*)user_data()); duke@435: } else { duke@435: res = invoke_primitive_field_callback_for_instance_fields(&wrapper, duke@435: obj, duke@435: cb, duke@435: (void*)user_data()); duke@435: } duke@435: if (check_flags_for_abort(res)) return; duke@435: } duke@435: duke@435: // string callback duke@435: if (!is_array && duke@435: callbacks()->string_primitive_value_callback != NULL && duke@435: obj->klass() == SystemDictionary::string_klass()) { duke@435: jint res = invoke_string_value_callback( duke@435: callbacks()->string_primitive_value_callback, duke@435: &wrapper, duke@435: obj, duke@435: (void*)user_data() ); duke@435: if (check_flags_for_abort(res)) return; duke@435: } duke@435: duke@435: // array callback duke@435: if (is_array && duke@435: callbacks()->array_primitive_value_callback != NULL && duke@435: obj->is_typeArray()) { duke@435: jint res = invoke_array_primitive_value_callback( duke@435: callbacks()->array_primitive_value_callback, duke@435: &wrapper, duke@435: obj, duke@435: (void*)user_data() ); duke@435: if (check_flags_for_abort(res)) return; duke@435: } duke@435: }; duke@435: duke@435: duke@435: // Deprecated function to iterate over all objects in the heap duke@435: void JvmtiTagMap::iterate_over_heap(jvmtiHeapObjectFilter object_filter, duke@435: KlassHandle klass, duke@435: jvmtiHeapObjectCallback heap_object_callback, duke@435: const void* user_data) duke@435: { duke@435: MutexLocker ml(Heap_lock); duke@435: IterateOverHeapObjectClosure blk(this, duke@435: klass, duke@435: object_filter, duke@435: heap_object_callback, duke@435: user_data); duke@435: VM_HeapIterateOperation op(&blk); duke@435: VMThread::execute(&op); duke@435: } duke@435: duke@435: duke@435: // Iterates over all objects in the heap duke@435: void JvmtiTagMap::iterate_through_heap(jint heap_filter, duke@435: KlassHandle klass, duke@435: const jvmtiHeapCallbacks* callbacks, duke@435: const void* user_data) duke@435: { duke@435: MutexLocker ml(Heap_lock); duke@435: IterateThroughHeapObjectClosure blk(this, duke@435: klass, duke@435: heap_filter, duke@435: callbacks, duke@435: user_data); duke@435: VM_HeapIterateOperation op(&blk); duke@435: VMThread::execute(&op); duke@435: } duke@435: duke@435: // support class for get_objects_with_tags duke@435: duke@435: class TagObjectCollector : public JvmtiTagHashmapEntryClosure { duke@435: private: duke@435: JvmtiEnv* _env; duke@435: jlong* _tags; duke@435: jint _tag_count; duke@435: duke@435: GrowableArray* _object_results; // collected objects (JNI weak refs) duke@435: GrowableArray* _tag_results; // collected tags duke@435: duke@435: public: duke@435: TagObjectCollector(JvmtiEnv* env, const jlong* tags, jint tag_count) { duke@435: _env = env; duke@435: _tags = (jlong*)tags; duke@435: _tag_count = tag_count; duke@435: _object_results = new (ResourceObj::C_HEAP) GrowableArray(1,true); duke@435: _tag_results = new (ResourceObj::C_HEAP) GrowableArray(1,true); duke@435: } duke@435: duke@435: ~TagObjectCollector() { duke@435: delete _object_results; duke@435: delete _tag_results; duke@435: } duke@435: duke@435: // for each tagged object check if the tag value matches duke@435: // - if it matches then we create a JNI local reference to the object duke@435: // and record the reference and tag value. duke@435: // duke@435: void do_entry(JvmtiTagHashmapEntry* entry) { duke@435: for (int i=0; i<_tag_count; i++) { duke@435: if (_tags[i] == entry->tag()) { duke@435: oop o = JNIHandles::resolve(entry->object()); duke@435: assert(o != NULL && o != JNIHandles::deleted_handle(), "sanity check"); duke@435: duke@435: // the mirror is tagged duke@435: if (o->is_klass()) { duke@435: klassOop k = (klassOop)o; duke@435: o = Klass::cast(k)->java_mirror(); duke@435: } duke@435: duke@435: jobject ref = JNIHandles::make_local(JavaThread::current(), o); duke@435: _object_results->append(ref); duke@435: _tag_results->append((uint64_t)entry->tag()); duke@435: } duke@435: } duke@435: } duke@435: duke@435: // return the results from the collection duke@435: // duke@435: jvmtiError result(jint* count_ptr, jobject** object_result_ptr, jlong** tag_result_ptr) { duke@435: jvmtiError error; duke@435: int count = _object_results->length(); duke@435: assert(count >= 0, "sanity check"); duke@435: duke@435: // if object_result_ptr is not NULL then allocate the result and copy duke@435: // in the object references. duke@435: if (object_result_ptr != NULL) { duke@435: error = _env->Allocate(count * sizeof(jobject), (unsigned char**)object_result_ptr); duke@435: if (error != JVMTI_ERROR_NONE) { duke@435: return error; duke@435: } duke@435: for (int i=0; iat(i); duke@435: } duke@435: } duke@435: duke@435: // if tag_result_ptr is not NULL then allocate the result and copy duke@435: // in the tag values. duke@435: if (tag_result_ptr != NULL) { duke@435: error = _env->Allocate(count * sizeof(jlong), (unsigned char**)tag_result_ptr); duke@435: if (error != JVMTI_ERROR_NONE) { duke@435: if (object_result_ptr != NULL) { duke@435: _env->Deallocate((unsigned char*)object_result_ptr); duke@435: } duke@435: return error; duke@435: } duke@435: for (int i=0; iat(i); duke@435: } duke@435: } duke@435: duke@435: *count_ptr = count; duke@435: return JVMTI_ERROR_NONE; duke@435: } duke@435: }; duke@435: duke@435: // return the list of objects with the specified tags duke@435: jvmtiError JvmtiTagMap::get_objects_with_tags(const jlong* tags, duke@435: jint count, jint* count_ptr, jobject** object_result_ptr, jlong** tag_result_ptr) { duke@435: duke@435: TagObjectCollector collector(env(), tags, count); duke@435: { duke@435: // iterate over all tagged objects duke@435: MutexLocker ml(lock()); duke@435: entry_iterate(&collector); duke@435: } duke@435: return collector.result(count_ptr, object_result_ptr, tag_result_ptr); duke@435: } duke@435: duke@435: duke@435: // ObjectMarker is used to support the marking objects when walking the duke@435: // heap. duke@435: // duke@435: // This implementation uses the existing mark bits in an object for duke@435: // marking. Objects that are marked must later have their headers restored. duke@435: // As most objects are unlocked and don't have their identity hash computed duke@435: // we don't have to save their headers. Instead we save the headers that duke@435: // are "interesting". Later when the headers are restored this implementation duke@435: // restores all headers to their initial value and then restores the few duke@435: // objects that had interesting headers. duke@435: // duke@435: // Future work: This implementation currently uses growable arrays to save duke@435: // the oop and header of interesting objects. As an optimization we could duke@435: // use the same technique as the GC and make use of the unused area duke@435: // between top() and end(). duke@435: // duke@435: duke@435: // An ObjectClosure used to restore the mark bits of an object duke@435: class RestoreMarksClosure : public ObjectClosure { duke@435: public: duke@435: void do_object(oop o) { duke@435: if (o != NULL) { duke@435: markOop mark = o->mark(); duke@435: if (mark->is_marked()) { duke@435: o->init_mark(); duke@435: } duke@435: } duke@435: } duke@435: }; duke@435: duke@435: // ObjectMarker provides the mark and visited functions duke@435: class ObjectMarker : AllStatic { duke@435: private: duke@435: // saved headers duke@435: static GrowableArray* _saved_oop_stack; duke@435: static GrowableArray* _saved_mark_stack; duke@435: duke@435: public: duke@435: static void init(); // initialize duke@435: static void done(); // clean-up duke@435: duke@435: static inline void mark(oop o); // mark an object duke@435: static inline bool visited(oop o); // check if object has been visited duke@435: }; duke@435: duke@435: GrowableArray* ObjectMarker::_saved_oop_stack = NULL; duke@435: GrowableArray* ObjectMarker::_saved_mark_stack = NULL; duke@435: duke@435: // initialize ObjectMarker - prepares for object marking duke@435: void ObjectMarker::init() { duke@435: assert(Thread::current()->is_VM_thread(), "must be VMThread"); duke@435: duke@435: // prepare heap for iteration duke@435: Universe::heap()->ensure_parsability(false); // no need to retire TLABs duke@435: duke@435: // create stacks for interesting headers duke@435: _saved_mark_stack = new (ResourceObj::C_HEAP) GrowableArray(4000, true); duke@435: _saved_oop_stack = new (ResourceObj::C_HEAP) GrowableArray(4000, true); duke@435: duke@435: if (UseBiasedLocking) { duke@435: BiasedLocking::preserve_marks(); duke@435: } duke@435: } duke@435: duke@435: // Object marking is done so restore object headers duke@435: void ObjectMarker::done() { duke@435: // iterate over all objects and restore the mark bits to duke@435: // their initial value duke@435: RestoreMarksClosure blk; duke@435: Universe::heap()->object_iterate(&blk); duke@435: duke@435: // When sharing is enabled we need to restore the headers of the objects duke@435: // in the readwrite space too. duke@435: if (UseSharedSpaces) { duke@435: GenCollectedHeap* gch = GenCollectedHeap::heap(); duke@435: CompactingPermGenGen* gen = (CompactingPermGenGen*)gch->perm_gen(); duke@435: gen->rw_space()->object_iterate(&blk); duke@435: } duke@435: duke@435: // now restore the interesting headers duke@435: for (int i = 0; i < _saved_oop_stack->length(); i++) { duke@435: oop o = _saved_oop_stack->at(i); duke@435: markOop mark = _saved_mark_stack->at(i); duke@435: o->set_mark(mark); duke@435: } duke@435: duke@435: if (UseBiasedLocking) { duke@435: BiasedLocking::restore_marks(); duke@435: } duke@435: duke@435: // free the stacks duke@435: delete _saved_oop_stack; duke@435: delete _saved_mark_stack; duke@435: } duke@435: duke@435: // mark an object duke@435: inline void ObjectMarker::mark(oop o) { duke@435: assert(Universe::heap()->is_in(o), "sanity check"); duke@435: assert(!o->mark()->is_marked(), "should only mark an object once"); duke@435: duke@435: // object's mark word duke@435: markOop mark = o->mark(); duke@435: duke@435: if (mark->must_be_preserved(o)) { duke@435: _saved_mark_stack->push(mark); duke@435: _saved_oop_stack->push(o); duke@435: } duke@435: duke@435: // mark the object duke@435: o->set_mark(markOopDesc::prototype()->set_marked()); duke@435: } duke@435: duke@435: // return true if object is marked duke@435: inline bool ObjectMarker::visited(oop o) { duke@435: return o->mark()->is_marked(); duke@435: } duke@435: duke@435: // Stack allocated class to help ensure that ObjectMarker is used duke@435: // correctly. Constructor initializes ObjectMarker, destructor calls duke@435: // ObjectMarker's done() function to restore object headers. duke@435: class ObjectMarkerController : public StackObj { duke@435: public: duke@435: ObjectMarkerController() { duke@435: ObjectMarker::init(); duke@435: } duke@435: ~ObjectMarkerController() { duke@435: ObjectMarker::done(); duke@435: } duke@435: }; duke@435: duke@435: duke@435: // helper to map a jvmtiHeapReferenceKind to an old style jvmtiHeapRootKind duke@435: // (not performance critical as only used for roots) duke@435: static jvmtiHeapRootKind toJvmtiHeapRootKind(jvmtiHeapReferenceKind kind) { duke@435: switch (kind) { duke@435: case JVMTI_HEAP_REFERENCE_JNI_GLOBAL: return JVMTI_HEAP_ROOT_JNI_GLOBAL; duke@435: case JVMTI_HEAP_REFERENCE_SYSTEM_CLASS: return JVMTI_HEAP_ROOT_SYSTEM_CLASS; duke@435: case JVMTI_HEAP_REFERENCE_MONITOR: return JVMTI_HEAP_ROOT_MONITOR; duke@435: case JVMTI_HEAP_REFERENCE_STACK_LOCAL: return JVMTI_HEAP_ROOT_STACK_LOCAL; duke@435: case JVMTI_HEAP_REFERENCE_JNI_LOCAL: return JVMTI_HEAP_ROOT_JNI_LOCAL; duke@435: case JVMTI_HEAP_REFERENCE_THREAD: return JVMTI_HEAP_ROOT_THREAD; duke@435: case JVMTI_HEAP_REFERENCE_OTHER: return JVMTI_HEAP_ROOT_OTHER; duke@435: default: ShouldNotReachHere(); return JVMTI_HEAP_ROOT_OTHER; duke@435: } duke@435: } duke@435: duke@435: // Base class for all heap walk contexts. The base class maintains a flag duke@435: // to indicate if the context is valid or not. duke@435: class HeapWalkContext VALUE_OBJ_CLASS_SPEC { duke@435: private: duke@435: bool _valid; duke@435: public: duke@435: HeapWalkContext(bool valid) { _valid = valid; } duke@435: void invalidate() { _valid = false; } duke@435: bool is_valid() const { return _valid; } duke@435: }; duke@435: duke@435: // A basic heap walk context for the deprecated heap walking functions. duke@435: // The context for a basic heap walk are the callbacks and fields used by duke@435: // the referrer caching scheme. duke@435: class BasicHeapWalkContext: public HeapWalkContext { duke@435: private: duke@435: jvmtiHeapRootCallback _heap_root_callback; duke@435: jvmtiStackReferenceCallback _stack_ref_callback; duke@435: jvmtiObjectReferenceCallback _object_ref_callback; duke@435: duke@435: // used for caching duke@435: oop _last_referrer; duke@435: jlong _last_referrer_tag; duke@435: duke@435: public: duke@435: BasicHeapWalkContext() : HeapWalkContext(false) { } duke@435: duke@435: BasicHeapWalkContext(jvmtiHeapRootCallback heap_root_callback, duke@435: jvmtiStackReferenceCallback stack_ref_callback, duke@435: jvmtiObjectReferenceCallback object_ref_callback) : duke@435: HeapWalkContext(true), duke@435: _heap_root_callback(heap_root_callback), duke@435: _stack_ref_callback(stack_ref_callback), duke@435: _object_ref_callback(object_ref_callback), duke@435: _last_referrer(NULL), duke@435: _last_referrer_tag(0) { duke@435: } duke@435: duke@435: // accessors duke@435: jvmtiHeapRootCallback heap_root_callback() const { return _heap_root_callback; } duke@435: jvmtiStackReferenceCallback stack_ref_callback() const { return _stack_ref_callback; } duke@435: jvmtiObjectReferenceCallback object_ref_callback() const { return _object_ref_callback; } duke@435: duke@435: oop last_referrer() const { return _last_referrer; } duke@435: void set_last_referrer(oop referrer) { _last_referrer = referrer; } duke@435: jlong last_referrer_tag() const { return _last_referrer_tag; } duke@435: void set_last_referrer_tag(jlong value) { _last_referrer_tag = value; } duke@435: }; duke@435: duke@435: // The advanced heap walk context for the FollowReferences functions. duke@435: // The context is the callbacks, and the fields used for filtering. duke@435: class AdvancedHeapWalkContext: public HeapWalkContext { duke@435: private: duke@435: jint _heap_filter; duke@435: KlassHandle _klass_filter; duke@435: const jvmtiHeapCallbacks* _heap_callbacks; duke@435: duke@435: public: duke@435: AdvancedHeapWalkContext() : HeapWalkContext(false) { } duke@435: duke@435: AdvancedHeapWalkContext(jint heap_filter, duke@435: KlassHandle klass_filter, duke@435: const jvmtiHeapCallbacks* heap_callbacks) : duke@435: HeapWalkContext(true), duke@435: _heap_filter(heap_filter), duke@435: _klass_filter(klass_filter), duke@435: _heap_callbacks(heap_callbacks) { duke@435: } duke@435: duke@435: // accessors duke@435: jint heap_filter() const { return _heap_filter; } duke@435: KlassHandle klass_filter() const { return _klass_filter; } duke@435: duke@435: const jvmtiHeapReferenceCallback heap_reference_callback() const { duke@435: return _heap_callbacks->heap_reference_callback; duke@435: }; duke@435: const jvmtiPrimitiveFieldCallback primitive_field_callback() const { duke@435: return _heap_callbacks->primitive_field_callback; duke@435: } duke@435: const jvmtiArrayPrimitiveValueCallback array_primitive_value_callback() const { duke@435: return _heap_callbacks->array_primitive_value_callback; duke@435: } duke@435: const jvmtiStringPrimitiveValueCallback string_primitive_value_callback() const { duke@435: return _heap_callbacks->string_primitive_value_callback; duke@435: } duke@435: }; duke@435: duke@435: // The CallbackInvoker is a class with static functions that the heap walk can call duke@435: // into to invoke callbacks. It works in one of two modes. The "basic" mode is duke@435: // used for the deprecated IterateOverReachableObjects functions. The "advanced" duke@435: // mode is for the newer FollowReferences function which supports a lot of duke@435: // additional callbacks. duke@435: class CallbackInvoker : AllStatic { duke@435: private: duke@435: // heap walk styles duke@435: enum { basic, advanced }; duke@435: static int _heap_walk_type; duke@435: static bool is_basic_heap_walk() { return _heap_walk_type == basic; } duke@435: static bool is_advanced_heap_walk() { return _heap_walk_type == advanced; } duke@435: duke@435: // context for basic style heap walk duke@435: static BasicHeapWalkContext _basic_context; duke@435: static BasicHeapWalkContext* basic_context() { duke@435: assert(_basic_context.is_valid(), "invalid"); duke@435: return &_basic_context; duke@435: } duke@435: duke@435: // context for advanced style heap walk duke@435: static AdvancedHeapWalkContext _advanced_context; duke@435: static AdvancedHeapWalkContext* advanced_context() { duke@435: assert(_advanced_context.is_valid(), "invalid"); duke@435: return &_advanced_context; duke@435: } duke@435: duke@435: // context needed for all heap walks duke@435: static JvmtiTagMap* _tag_map; duke@435: static const void* _user_data; duke@435: static GrowableArray* _visit_stack; duke@435: duke@435: // accessors duke@435: static JvmtiTagMap* tag_map() { return _tag_map; } duke@435: static const void* user_data() { return _user_data; } duke@435: static GrowableArray* visit_stack() { return _visit_stack; } duke@435: duke@435: // if the object hasn't been visited then push it onto the visit stack duke@435: // so that it will be visited later duke@435: static inline bool check_for_visit(oop obj) { duke@435: if (!ObjectMarker::visited(obj)) visit_stack()->push(obj); duke@435: return true; duke@435: } duke@435: duke@435: // invoke basic style callbacks duke@435: static inline bool invoke_basic_heap_root_callback duke@435: (jvmtiHeapRootKind root_kind, oop obj); duke@435: static inline bool invoke_basic_stack_ref_callback duke@435: (jvmtiHeapRootKind root_kind, jlong thread_tag, jint depth, jmethodID method, duke@435: int slot, oop obj); duke@435: static inline bool invoke_basic_object_reference_callback duke@435: (jvmtiObjectReferenceKind ref_kind, oop referrer, oop referree, jint index); duke@435: duke@435: // invoke advanced style callbacks duke@435: static inline bool invoke_advanced_heap_root_callback duke@435: (jvmtiHeapReferenceKind ref_kind, oop obj); duke@435: static inline bool invoke_advanced_stack_ref_callback duke@435: (jvmtiHeapReferenceKind ref_kind, jlong thread_tag, jlong tid, int depth, duke@435: jmethodID method, jlocation bci, jint slot, oop obj); duke@435: static inline bool invoke_advanced_object_reference_callback duke@435: (jvmtiHeapReferenceKind ref_kind, oop referrer, oop referree, jint index); duke@435: duke@435: // used to report the value of primitive fields duke@435: static inline bool report_primitive_field duke@435: (jvmtiHeapReferenceKind ref_kind, oop obj, jint index, address addr, char type); duke@435: duke@435: public: duke@435: // initialize for basic mode duke@435: static void initialize_for_basic_heap_walk(JvmtiTagMap* tag_map, duke@435: GrowableArray* visit_stack, duke@435: const void* user_data, duke@435: BasicHeapWalkContext context); duke@435: duke@435: // initialize for advanced mode duke@435: static void initialize_for_advanced_heap_walk(JvmtiTagMap* tag_map, duke@435: GrowableArray* visit_stack, duke@435: const void* user_data, duke@435: AdvancedHeapWalkContext context); duke@435: duke@435: // functions to report roots duke@435: static inline bool report_simple_root(jvmtiHeapReferenceKind kind, oop o); duke@435: static inline bool report_jni_local_root(jlong thread_tag, jlong tid, jint depth, duke@435: jmethodID m, oop o); duke@435: static inline bool report_stack_ref_root(jlong thread_tag, jlong tid, jint depth, duke@435: jmethodID method, jlocation bci, jint slot, oop o); duke@435: duke@435: // functions to report references duke@435: static inline bool report_array_element_reference(oop referrer, oop referree, jint index); duke@435: static inline bool report_class_reference(oop referrer, oop referree); duke@435: static inline bool report_class_loader_reference(oop referrer, oop referree); duke@435: static inline bool report_signers_reference(oop referrer, oop referree); duke@435: static inline bool report_protection_domain_reference(oop referrer, oop referree); duke@435: static inline bool report_superclass_reference(oop referrer, oop referree); duke@435: static inline bool report_interface_reference(oop referrer, oop referree); duke@435: static inline bool report_static_field_reference(oop referrer, oop referree, jint slot); duke@435: static inline bool report_field_reference(oop referrer, oop referree, jint slot); duke@435: static inline bool report_constant_pool_reference(oop referrer, oop referree, jint index); duke@435: static inline bool report_primitive_array_values(oop array); duke@435: static inline bool report_string_value(oop str); duke@435: static inline bool report_primitive_instance_field(oop o, jint index, address value, char type); duke@435: static inline bool report_primitive_static_field(oop o, jint index, address value, char type); duke@435: }; duke@435: duke@435: // statics duke@435: int CallbackInvoker::_heap_walk_type; duke@435: BasicHeapWalkContext CallbackInvoker::_basic_context; duke@435: AdvancedHeapWalkContext CallbackInvoker::_advanced_context; duke@435: JvmtiTagMap* CallbackInvoker::_tag_map; duke@435: const void* CallbackInvoker::_user_data; duke@435: GrowableArray* CallbackInvoker::_visit_stack; duke@435: duke@435: // initialize for basic heap walk (IterateOverReachableObjects et al) duke@435: void CallbackInvoker::initialize_for_basic_heap_walk(JvmtiTagMap* tag_map, duke@435: GrowableArray* visit_stack, duke@435: const void* user_data, duke@435: BasicHeapWalkContext context) { duke@435: _tag_map = tag_map; duke@435: _visit_stack = visit_stack; duke@435: _user_data = user_data; duke@435: _basic_context = context; duke@435: _advanced_context.invalidate(); // will trigger assertion if used duke@435: _heap_walk_type = basic; duke@435: } duke@435: duke@435: // initialize for advanced heap walk (FollowReferences) duke@435: void CallbackInvoker::initialize_for_advanced_heap_walk(JvmtiTagMap* tag_map, duke@435: GrowableArray* visit_stack, duke@435: const void* user_data, duke@435: AdvancedHeapWalkContext context) { duke@435: _tag_map = tag_map; duke@435: _visit_stack = visit_stack; duke@435: _user_data = user_data; duke@435: _advanced_context = context; duke@435: _basic_context.invalidate(); // will trigger assertion if used duke@435: _heap_walk_type = advanced; duke@435: } duke@435: duke@435: duke@435: // invoke basic style heap root callback duke@435: inline bool CallbackInvoker::invoke_basic_heap_root_callback(jvmtiHeapRootKind root_kind, oop obj) { duke@435: assert(ServiceUtil::visible_oop(obj), "checking"); duke@435: duke@435: // if we heap roots should be reported duke@435: jvmtiHeapRootCallback cb = basic_context()->heap_root_callback(); duke@435: if (cb == NULL) { duke@435: return check_for_visit(obj); duke@435: } duke@435: duke@435: CallbackWrapper wrapper(tag_map(), obj); duke@435: jvmtiIterationControl control = (*cb)(root_kind, duke@435: wrapper.klass_tag(), duke@435: wrapper.obj_size(), duke@435: wrapper.obj_tag_p(), duke@435: (void*)user_data()); duke@435: // push root to visit stack when following references duke@435: if (control == JVMTI_ITERATION_CONTINUE && duke@435: basic_context()->object_ref_callback() != NULL) { duke@435: visit_stack()->push(obj); duke@435: } duke@435: return control != JVMTI_ITERATION_ABORT; duke@435: } duke@435: duke@435: // invoke basic style stack ref callback duke@435: inline bool CallbackInvoker::invoke_basic_stack_ref_callback(jvmtiHeapRootKind root_kind, duke@435: jlong thread_tag, duke@435: jint depth, duke@435: jmethodID method, duke@435: jint slot, duke@435: oop obj) { duke@435: assert(ServiceUtil::visible_oop(obj), "checking"); duke@435: duke@435: // if we stack refs should be reported duke@435: jvmtiStackReferenceCallback cb = basic_context()->stack_ref_callback(); duke@435: if (cb == NULL) { duke@435: return check_for_visit(obj); duke@435: } duke@435: duke@435: CallbackWrapper wrapper(tag_map(), obj); duke@435: jvmtiIterationControl control = (*cb)(root_kind, duke@435: wrapper.klass_tag(), duke@435: wrapper.obj_size(), duke@435: wrapper.obj_tag_p(), duke@435: thread_tag, duke@435: depth, duke@435: method, duke@435: slot, duke@435: (void*)user_data()); duke@435: // push root to visit stack when following references duke@435: if (control == JVMTI_ITERATION_CONTINUE && duke@435: basic_context()->object_ref_callback() != NULL) { duke@435: visit_stack()->push(obj); duke@435: } duke@435: return control != JVMTI_ITERATION_ABORT; duke@435: } duke@435: duke@435: // invoke basic style object reference callback duke@435: inline bool CallbackInvoker::invoke_basic_object_reference_callback(jvmtiObjectReferenceKind ref_kind, duke@435: oop referrer, duke@435: oop referree, duke@435: jint index) { duke@435: duke@435: assert(ServiceUtil::visible_oop(referrer), "checking"); duke@435: assert(ServiceUtil::visible_oop(referree), "checking"); duke@435: duke@435: BasicHeapWalkContext* context = basic_context(); duke@435: duke@435: // callback requires the referrer's tag. If it's the same referrer duke@435: // as the last call then we use the cached value. duke@435: jlong referrer_tag; duke@435: if (referrer == context->last_referrer()) { duke@435: referrer_tag = context->last_referrer_tag(); duke@435: } else { duke@435: referrer_tag = tag_for(tag_map(), klassOop_if_java_lang_Class(referrer)); duke@435: } duke@435: duke@435: // do the callback duke@435: CallbackWrapper wrapper(tag_map(), referree); duke@435: jvmtiObjectReferenceCallback cb = context->object_ref_callback(); duke@435: jvmtiIterationControl control = (*cb)(ref_kind, duke@435: wrapper.klass_tag(), duke@435: wrapper.obj_size(), duke@435: wrapper.obj_tag_p(), duke@435: referrer_tag, duke@435: index, duke@435: (void*)user_data()); duke@435: duke@435: // record referrer and referrer tag. For self-references record the duke@435: // tag value from the callback as this might differ from referrer_tag. duke@435: context->set_last_referrer(referrer); duke@435: if (referrer == referree) { duke@435: context->set_last_referrer_tag(*wrapper.obj_tag_p()); duke@435: } else { duke@435: context->set_last_referrer_tag(referrer_tag); duke@435: } duke@435: duke@435: if (control == JVMTI_ITERATION_CONTINUE) { duke@435: return check_for_visit(referree); duke@435: } else { duke@435: return control != JVMTI_ITERATION_ABORT; duke@435: } duke@435: } duke@435: duke@435: // invoke advanced style heap root callback duke@435: inline bool CallbackInvoker::invoke_advanced_heap_root_callback(jvmtiHeapReferenceKind ref_kind, duke@435: oop obj) { duke@435: assert(ServiceUtil::visible_oop(obj), "checking"); duke@435: duke@435: AdvancedHeapWalkContext* context = advanced_context(); duke@435: duke@435: // check that callback is provided duke@435: jvmtiHeapReferenceCallback cb = context->heap_reference_callback(); duke@435: if (cb == NULL) { duke@435: return check_for_visit(obj); duke@435: } duke@435: duke@435: // apply class filter duke@435: if (is_filtered_by_klass_filter(obj, context->klass_filter())) { duke@435: return check_for_visit(obj); duke@435: } duke@435: duke@435: // setup the callback wrapper duke@435: CallbackWrapper wrapper(tag_map(), obj); duke@435: duke@435: // apply tag filter duke@435: if (is_filtered_by_heap_filter(wrapper.obj_tag(), duke@435: wrapper.klass_tag(), duke@435: context->heap_filter())) { duke@435: return check_for_visit(obj); duke@435: } duke@435: duke@435: // for arrays we need the length, otherwise -1 duke@435: jint len = (jint)(obj->is_array() ? arrayOop(obj)->length() : -1); duke@435: duke@435: // invoke the callback duke@435: jint res = (*cb)(ref_kind, duke@435: NULL, // referrer info duke@435: wrapper.klass_tag(), duke@435: 0, // referrer_class_tag is 0 for heap root duke@435: wrapper.obj_size(), duke@435: wrapper.obj_tag_p(), duke@435: NULL, // referrer_tag_p duke@435: len, duke@435: (void*)user_data()); duke@435: if (res & JVMTI_VISIT_ABORT) { duke@435: return false;// referrer class tag duke@435: } duke@435: if (res & JVMTI_VISIT_OBJECTS) { duke@435: check_for_visit(obj); duke@435: } duke@435: return true; duke@435: } duke@435: duke@435: // report a reference from a thread stack to an object duke@435: inline bool CallbackInvoker::invoke_advanced_stack_ref_callback(jvmtiHeapReferenceKind ref_kind, duke@435: jlong thread_tag, duke@435: jlong tid, duke@435: int depth, duke@435: jmethodID method, duke@435: jlocation bci, duke@435: jint slot, duke@435: oop obj) { duke@435: assert(ServiceUtil::visible_oop(obj), "checking"); duke@435: duke@435: AdvancedHeapWalkContext* context = advanced_context(); duke@435: duke@435: // check that callback is provider duke@435: jvmtiHeapReferenceCallback cb = context->heap_reference_callback(); duke@435: if (cb == NULL) { duke@435: return check_for_visit(obj); duke@435: } duke@435: duke@435: // apply class filter duke@435: if (is_filtered_by_klass_filter(obj, context->klass_filter())) { duke@435: return check_for_visit(obj); duke@435: } duke@435: duke@435: // setup the callback wrapper duke@435: CallbackWrapper wrapper(tag_map(), obj); duke@435: duke@435: // apply tag filter duke@435: if (is_filtered_by_heap_filter(wrapper.obj_tag(), duke@435: wrapper.klass_tag(), duke@435: context->heap_filter())) { duke@435: return check_for_visit(obj); duke@435: } duke@435: duke@435: // setup the referrer info duke@435: jvmtiHeapReferenceInfo reference_info; duke@435: reference_info.stack_local.thread_tag = thread_tag; duke@435: reference_info.stack_local.thread_id = tid; duke@435: reference_info.stack_local.depth = depth; duke@435: reference_info.stack_local.method = method; duke@435: reference_info.stack_local.location = bci; duke@435: reference_info.stack_local.slot = slot; duke@435: duke@435: // for arrays we need the length, otherwise -1 duke@435: jint len = (jint)(obj->is_array() ? arrayOop(obj)->length() : -1); duke@435: duke@435: // call into the agent duke@435: int res = (*cb)(ref_kind, duke@435: &reference_info, duke@435: wrapper.klass_tag(), duke@435: 0, // referrer_class_tag is 0 for heap root (stack) duke@435: wrapper.obj_size(), duke@435: wrapper.obj_tag_p(), duke@435: NULL, // referrer_tag is 0 for root duke@435: len, duke@435: (void*)user_data()); duke@435: duke@435: if (res & JVMTI_VISIT_ABORT) { duke@435: return false; duke@435: } duke@435: if (res & JVMTI_VISIT_OBJECTS) { duke@435: check_for_visit(obj); duke@435: } duke@435: return true; duke@435: } duke@435: duke@435: // This mask is used to pass reference_info to a jvmtiHeapReferenceCallback duke@435: // only for ref_kinds defined by the JVM TI spec. Otherwise, NULL is passed. duke@435: #define REF_INFO_MASK ((1 << JVMTI_HEAP_REFERENCE_FIELD) \ duke@435: | (1 << JVMTI_HEAP_REFERENCE_STATIC_FIELD) \ duke@435: | (1 << JVMTI_HEAP_REFERENCE_ARRAY_ELEMENT) \ duke@435: | (1 << JVMTI_HEAP_REFERENCE_CONSTANT_POOL) \ duke@435: | (1 << JVMTI_HEAP_REFERENCE_STACK_LOCAL) \ duke@435: | (1 << JVMTI_HEAP_REFERENCE_JNI_LOCAL)) duke@435: duke@435: // invoke the object reference callback to report a reference duke@435: inline bool CallbackInvoker::invoke_advanced_object_reference_callback(jvmtiHeapReferenceKind ref_kind, duke@435: oop referrer, duke@435: oop obj, duke@435: jint index) duke@435: { duke@435: // field index is only valid field in reference_info duke@435: static jvmtiHeapReferenceInfo reference_info = { 0 }; duke@435: duke@435: assert(ServiceUtil::visible_oop(referrer), "checking"); duke@435: assert(ServiceUtil::visible_oop(obj), "checking"); duke@435: duke@435: AdvancedHeapWalkContext* context = advanced_context(); duke@435: duke@435: // check that callback is provider duke@435: jvmtiHeapReferenceCallback cb = context->heap_reference_callback(); duke@435: if (cb == NULL) { duke@435: return check_for_visit(obj); duke@435: } duke@435: duke@435: // apply class filter duke@435: if (is_filtered_by_klass_filter(obj, context->klass_filter())) { duke@435: return check_for_visit(obj); duke@435: } duke@435: duke@435: // setup the callback wrapper duke@435: TwoOopCallbackWrapper wrapper(tag_map(), referrer, obj); duke@435: duke@435: // apply tag filter duke@435: if (is_filtered_by_heap_filter(wrapper.obj_tag(), duke@435: wrapper.klass_tag(), duke@435: context->heap_filter())) { duke@435: return check_for_visit(obj); duke@435: } duke@435: duke@435: // field index is only valid field in reference_info duke@435: reference_info.field.index = index; duke@435: duke@435: // for arrays we need the length, otherwise -1 duke@435: jint len = (jint)(obj->is_array() ? arrayOop(obj)->length() : -1); duke@435: duke@435: // invoke the callback duke@435: int res = (*cb)(ref_kind, duke@435: (REF_INFO_MASK & (1 << ref_kind)) ? &reference_info : NULL, duke@435: wrapper.klass_tag(), duke@435: wrapper.referrer_klass_tag(), duke@435: wrapper.obj_size(), duke@435: wrapper.obj_tag_p(), duke@435: wrapper.referrer_tag_p(), duke@435: len, duke@435: (void*)user_data()); duke@435: duke@435: if (res & JVMTI_VISIT_ABORT) { duke@435: return false; duke@435: } duke@435: if (res & JVMTI_VISIT_OBJECTS) { duke@435: check_for_visit(obj); duke@435: } duke@435: return true; duke@435: } duke@435: duke@435: // report a "simple root" duke@435: inline bool CallbackInvoker::report_simple_root(jvmtiHeapReferenceKind kind, oop obj) { duke@435: assert(kind != JVMTI_HEAP_REFERENCE_STACK_LOCAL && duke@435: kind != JVMTI_HEAP_REFERENCE_JNI_LOCAL, "not a simple root"); duke@435: assert(ServiceUtil::visible_oop(obj), "checking"); duke@435: duke@435: if (is_basic_heap_walk()) { duke@435: // map to old style root kind duke@435: jvmtiHeapRootKind root_kind = toJvmtiHeapRootKind(kind); duke@435: return invoke_basic_heap_root_callback(root_kind, obj); duke@435: } else { duke@435: assert(is_advanced_heap_walk(), "wrong heap walk type"); duke@435: return invoke_advanced_heap_root_callback(kind, obj); duke@435: } duke@435: } duke@435: duke@435: duke@435: // invoke the primitive array values duke@435: inline bool CallbackInvoker::report_primitive_array_values(oop obj) { duke@435: assert(obj->is_typeArray(), "not a primitive array"); duke@435: duke@435: AdvancedHeapWalkContext* context = advanced_context(); duke@435: assert(context->array_primitive_value_callback() != NULL, "no callback"); duke@435: duke@435: // apply class filter duke@435: if (is_filtered_by_klass_filter(obj, context->klass_filter())) { duke@435: return true; duke@435: } duke@435: duke@435: CallbackWrapper wrapper(tag_map(), obj); duke@435: duke@435: // apply tag filter duke@435: if (is_filtered_by_heap_filter(wrapper.obj_tag(), duke@435: wrapper.klass_tag(), duke@435: context->heap_filter())) { duke@435: return true; duke@435: } duke@435: duke@435: // invoke the callback duke@435: int res = invoke_array_primitive_value_callback(context->array_primitive_value_callback(), duke@435: &wrapper, duke@435: obj, duke@435: (void*)user_data()); duke@435: return (!(res & JVMTI_VISIT_ABORT)); duke@435: } duke@435: duke@435: // invoke the string value callback duke@435: inline bool CallbackInvoker::report_string_value(oop str) { duke@435: assert(str->klass() == SystemDictionary::string_klass(), "not a string"); duke@435: duke@435: AdvancedHeapWalkContext* context = advanced_context(); duke@435: assert(context->string_primitive_value_callback() != NULL, "no callback"); duke@435: duke@435: // apply class filter duke@435: if (is_filtered_by_klass_filter(str, context->klass_filter())) { duke@435: return true; duke@435: } duke@435: duke@435: CallbackWrapper wrapper(tag_map(), str); duke@435: duke@435: // apply tag filter duke@435: if (is_filtered_by_heap_filter(wrapper.obj_tag(), duke@435: wrapper.klass_tag(), duke@435: context->heap_filter())) { duke@435: return true; duke@435: } duke@435: duke@435: // invoke the callback duke@435: int res = invoke_string_value_callback(context->string_primitive_value_callback(), duke@435: &wrapper, duke@435: str, duke@435: (void*)user_data()); duke@435: return (!(res & JVMTI_VISIT_ABORT)); duke@435: } duke@435: duke@435: // invoke the primitive field callback duke@435: inline bool CallbackInvoker::report_primitive_field(jvmtiHeapReferenceKind ref_kind, duke@435: oop obj, duke@435: jint index, duke@435: address addr, duke@435: char type) duke@435: { duke@435: // for primitive fields only the index will be set duke@435: static jvmtiHeapReferenceInfo reference_info = { 0 }; duke@435: duke@435: AdvancedHeapWalkContext* context = advanced_context(); duke@435: assert(context->primitive_field_callback() != NULL, "no callback"); duke@435: duke@435: // apply class filter duke@435: if (is_filtered_by_klass_filter(obj, context->klass_filter())) { duke@435: return true; duke@435: } duke@435: duke@435: CallbackWrapper wrapper(tag_map(), obj); duke@435: duke@435: // apply tag filter duke@435: if (is_filtered_by_heap_filter(wrapper.obj_tag(), duke@435: wrapper.klass_tag(), duke@435: context->heap_filter())) { duke@435: return true; duke@435: } duke@435: duke@435: // the field index in the referrer duke@435: reference_info.field.index = index; duke@435: duke@435: // map the type duke@435: jvmtiPrimitiveType value_type = (jvmtiPrimitiveType)type; duke@435: duke@435: // setup the jvalue duke@435: jvalue value; duke@435: copy_to_jvalue(&value, addr, value_type); duke@435: duke@435: jvmtiPrimitiveFieldCallback cb = context->primitive_field_callback(); duke@435: int res = (*cb)(ref_kind, duke@435: &reference_info, duke@435: wrapper.klass_tag(), duke@435: wrapper.obj_tag_p(), duke@435: value, duke@435: value_type, duke@435: (void*)user_data()); duke@435: return (!(res & JVMTI_VISIT_ABORT)); duke@435: } duke@435: duke@435: duke@435: // instance field duke@435: inline bool CallbackInvoker::report_primitive_instance_field(oop obj, duke@435: jint index, duke@435: address value, duke@435: char type) { duke@435: return report_primitive_field(JVMTI_HEAP_REFERENCE_FIELD, duke@435: obj, duke@435: index, duke@435: value, duke@435: type); duke@435: } duke@435: duke@435: // static field duke@435: inline bool CallbackInvoker::report_primitive_static_field(oop obj, duke@435: jint index, duke@435: address value, duke@435: char type) { duke@435: return report_primitive_field(JVMTI_HEAP_REFERENCE_STATIC_FIELD, duke@435: obj, duke@435: index, duke@435: value, duke@435: type); duke@435: } duke@435: duke@435: // report a JNI local (root object) to the profiler duke@435: inline bool CallbackInvoker::report_jni_local_root(jlong thread_tag, jlong tid, jint depth, jmethodID m, oop obj) { duke@435: if (is_basic_heap_walk()) { duke@435: return invoke_basic_stack_ref_callback(JVMTI_HEAP_ROOT_JNI_LOCAL, duke@435: thread_tag, duke@435: depth, duke@435: m, duke@435: -1, duke@435: obj); duke@435: } else { duke@435: return invoke_advanced_stack_ref_callback(JVMTI_HEAP_REFERENCE_JNI_LOCAL, duke@435: thread_tag, tid, duke@435: depth, duke@435: m, duke@435: (jlocation)-1, duke@435: -1, duke@435: obj); duke@435: } duke@435: } duke@435: duke@435: duke@435: // report a local (stack reference, root object) duke@435: inline bool CallbackInvoker::report_stack_ref_root(jlong thread_tag, duke@435: jlong tid, duke@435: jint depth, duke@435: jmethodID method, duke@435: jlocation bci, duke@435: jint slot, duke@435: oop obj) { duke@435: if (is_basic_heap_walk()) { duke@435: return invoke_basic_stack_ref_callback(JVMTI_HEAP_ROOT_STACK_LOCAL, duke@435: thread_tag, duke@435: depth, duke@435: method, duke@435: slot, duke@435: obj); duke@435: } else { duke@435: return invoke_advanced_stack_ref_callback(JVMTI_HEAP_REFERENCE_STACK_LOCAL, duke@435: thread_tag, duke@435: tid, duke@435: depth, duke@435: method, duke@435: bci, duke@435: slot, duke@435: obj); duke@435: } duke@435: } duke@435: duke@435: // report an object referencing a class. duke@435: inline bool CallbackInvoker::report_class_reference(oop referrer, oop referree) { duke@435: if (is_basic_heap_walk()) { duke@435: return invoke_basic_object_reference_callback(JVMTI_REFERENCE_CLASS, referrer, referree, -1); duke@435: } else { duke@435: return invoke_advanced_object_reference_callback(JVMTI_HEAP_REFERENCE_CLASS, referrer, referree, -1); duke@435: } duke@435: } duke@435: duke@435: // report a class referencing its class loader. duke@435: inline bool CallbackInvoker::report_class_loader_reference(oop referrer, oop referree) { duke@435: if (is_basic_heap_walk()) { duke@435: return invoke_basic_object_reference_callback(JVMTI_REFERENCE_CLASS_LOADER, referrer, referree, -1); duke@435: } else { duke@435: return invoke_advanced_object_reference_callback(JVMTI_HEAP_REFERENCE_CLASS_LOADER, referrer, referree, -1); duke@435: } duke@435: } duke@435: duke@435: // report a class referencing its signers. duke@435: inline bool CallbackInvoker::report_signers_reference(oop referrer, oop referree) { duke@435: if (is_basic_heap_walk()) { duke@435: return invoke_basic_object_reference_callback(JVMTI_REFERENCE_SIGNERS, referrer, referree, -1); duke@435: } else { duke@435: return invoke_advanced_object_reference_callback(JVMTI_HEAP_REFERENCE_SIGNERS, referrer, referree, -1); duke@435: } duke@435: } duke@435: duke@435: // report a class referencing its protection domain.. duke@435: inline bool CallbackInvoker::report_protection_domain_reference(oop referrer, oop referree) { duke@435: if (is_basic_heap_walk()) { duke@435: return invoke_basic_object_reference_callback(JVMTI_REFERENCE_PROTECTION_DOMAIN, referrer, referree, -1); duke@435: } else { duke@435: return invoke_advanced_object_reference_callback(JVMTI_HEAP_REFERENCE_PROTECTION_DOMAIN, referrer, referree, -1); duke@435: } duke@435: } duke@435: duke@435: // report a class referencing its superclass. duke@435: inline bool CallbackInvoker::report_superclass_reference(oop referrer, oop referree) { duke@435: if (is_basic_heap_walk()) { duke@435: // Send this to be consistent with past implementation duke@435: return invoke_basic_object_reference_callback(JVMTI_REFERENCE_CLASS, referrer, referree, -1); duke@435: } else { duke@435: return invoke_advanced_object_reference_callback(JVMTI_HEAP_REFERENCE_SUPERCLASS, referrer, referree, -1); duke@435: } duke@435: } duke@435: duke@435: // report a class referencing one of its interfaces. duke@435: inline bool CallbackInvoker::report_interface_reference(oop referrer, oop referree) { duke@435: if (is_basic_heap_walk()) { duke@435: return invoke_basic_object_reference_callback(JVMTI_REFERENCE_INTERFACE, referrer, referree, -1); duke@435: } else { duke@435: return invoke_advanced_object_reference_callback(JVMTI_HEAP_REFERENCE_INTERFACE, referrer, referree, -1); duke@435: } duke@435: } duke@435: duke@435: // report a class referencing one of its static fields. duke@435: inline bool CallbackInvoker::report_static_field_reference(oop referrer, oop referree, jint slot) { duke@435: if (is_basic_heap_walk()) { duke@435: return invoke_basic_object_reference_callback(JVMTI_REFERENCE_STATIC_FIELD, referrer, referree, slot); duke@435: } else { duke@435: return invoke_advanced_object_reference_callback(JVMTI_HEAP_REFERENCE_STATIC_FIELD, referrer, referree, slot); duke@435: } duke@435: } duke@435: duke@435: // report an array referencing an element object duke@435: inline bool CallbackInvoker::report_array_element_reference(oop referrer, oop referree, jint index) { duke@435: if (is_basic_heap_walk()) { duke@435: return invoke_basic_object_reference_callback(JVMTI_REFERENCE_ARRAY_ELEMENT, referrer, referree, index); duke@435: } else { duke@435: return invoke_advanced_object_reference_callback(JVMTI_HEAP_REFERENCE_ARRAY_ELEMENT, referrer, referree, index); duke@435: } duke@435: } duke@435: duke@435: // report an object referencing an instance field object duke@435: inline bool CallbackInvoker::report_field_reference(oop referrer, oop referree, jint slot) { duke@435: if (is_basic_heap_walk()) { duke@435: return invoke_basic_object_reference_callback(JVMTI_REFERENCE_FIELD, referrer, referree, slot); duke@435: } else { duke@435: return invoke_advanced_object_reference_callback(JVMTI_HEAP_REFERENCE_FIELD, referrer, referree, slot); duke@435: } duke@435: } duke@435: duke@435: // report an array referencing an element object duke@435: inline bool CallbackInvoker::report_constant_pool_reference(oop referrer, oop referree, jint index) { duke@435: if (is_basic_heap_walk()) { duke@435: return invoke_basic_object_reference_callback(JVMTI_REFERENCE_CONSTANT_POOL, referrer, referree, index); duke@435: } else { duke@435: return invoke_advanced_object_reference_callback(JVMTI_HEAP_REFERENCE_CONSTANT_POOL, referrer, referree, index); duke@435: } duke@435: } duke@435: duke@435: // A supporting closure used to process simple roots duke@435: class SimpleRootsClosure : public OopClosure { duke@435: private: duke@435: jvmtiHeapReferenceKind _kind; duke@435: bool _continue; duke@435: duke@435: jvmtiHeapReferenceKind root_kind() { return _kind; } duke@435: duke@435: public: duke@435: void set_kind(jvmtiHeapReferenceKind kind) { duke@435: _kind = kind; duke@435: _continue = true; duke@435: } duke@435: duke@435: inline bool stopped() { duke@435: return !_continue; duke@435: } duke@435: duke@435: void do_oop(oop* obj_p) { duke@435: // iteration has terminated duke@435: if (stopped()) { duke@435: return; duke@435: } duke@435: duke@435: // ignore null or deleted handles duke@435: oop o = *obj_p; duke@435: if (o == NULL || o == JNIHandles::deleted_handle()) { duke@435: return; duke@435: } duke@435: duke@435: jvmtiHeapReferenceKind kind = root_kind(); duke@435: duke@435: // many roots are Klasses so we use the java mirror duke@435: if (o->is_klass()) { duke@435: klassOop k = (klassOop)o; duke@435: o = Klass::cast(k)->java_mirror(); duke@435: } else { duke@435: duke@435: // SystemDictionary::always_strong_oops_do reports the application duke@435: // class loader as a root. We want this root to be reported as duke@435: // a root kind of "OTHER" rather than "SYSTEM_CLASS". duke@435: if (o->is_instance() && root_kind() == JVMTI_HEAP_REFERENCE_SYSTEM_CLASS) { duke@435: kind = JVMTI_HEAP_REFERENCE_OTHER; duke@435: } duke@435: } duke@435: duke@435: // some objects are ignored - in the case of simple duke@435: // roots it's mostly symbolOops that we are skipping duke@435: // here. duke@435: if (!ServiceUtil::visible_oop(o)) { duke@435: return; duke@435: } duke@435: duke@435: // invoke the callback duke@435: _continue = CallbackInvoker::report_simple_root(kind, o); duke@435: duke@435: } coleenp@548: virtual void do_oop(narrowOop* obj_p) { ShouldNotReachHere(); } duke@435: }; duke@435: duke@435: // A supporting closure used to process JNI locals duke@435: class JNILocalRootsClosure : public OopClosure { duke@435: private: duke@435: jlong _thread_tag; duke@435: jlong _tid; duke@435: jint _depth; duke@435: jmethodID _method; duke@435: bool _continue; duke@435: public: duke@435: void set_context(jlong thread_tag, jlong tid, jint depth, jmethodID method) { duke@435: _thread_tag = thread_tag; duke@435: _tid = tid; duke@435: _depth = depth; duke@435: _method = method; duke@435: _continue = true; duke@435: } duke@435: duke@435: inline bool stopped() { duke@435: return !_continue; duke@435: } duke@435: duke@435: void do_oop(oop* obj_p) { duke@435: // iteration has terminated duke@435: if (stopped()) { duke@435: return; duke@435: } duke@435: duke@435: // ignore null or deleted handles duke@435: oop o = *obj_p; duke@435: if (o == NULL || o == JNIHandles::deleted_handle()) { duke@435: return; duke@435: } duke@435: duke@435: if (!ServiceUtil::visible_oop(o)) { duke@435: return; duke@435: } duke@435: duke@435: // invoke the callback duke@435: _continue = CallbackInvoker::report_jni_local_root(_thread_tag, _tid, _depth, _method, o); duke@435: } coleenp@548: virtual void do_oop(narrowOop* obj_p) { ShouldNotReachHere(); } duke@435: }; duke@435: duke@435: duke@435: // A VM operation to iterate over objects that are reachable from duke@435: // a set of roots or an initial object. duke@435: // duke@435: // For VM_HeapWalkOperation the set of roots used is :- duke@435: // duke@435: // - All JNI global references duke@435: // - All inflated monitors duke@435: // - All classes loaded by the boot class loader (or all classes duke@435: // in the event that class unloading is disabled) duke@435: // - All java threads duke@435: // - For each java thread then all locals and JNI local references duke@435: // on the thread's execution stack duke@435: // - All visible/explainable objects from Universes::oops_do duke@435: // duke@435: class VM_HeapWalkOperation: public VM_Operation { duke@435: private: duke@435: enum { duke@435: initial_visit_stack_size = 4000 duke@435: }; duke@435: duke@435: bool _is_advanced_heap_walk; // indicates FollowReferences duke@435: JvmtiTagMap* _tag_map; duke@435: Handle _initial_object; duke@435: GrowableArray* _visit_stack; // the visit stack duke@435: duke@435: bool _collecting_heap_roots; // are we collecting roots duke@435: bool _following_object_refs; // are we following object references duke@435: duke@435: bool _reporting_primitive_fields; // optional reporting duke@435: bool _reporting_primitive_array_values; duke@435: bool _reporting_string_values; duke@435: duke@435: GrowableArray* create_visit_stack() { duke@435: return new (ResourceObj::C_HEAP) GrowableArray(initial_visit_stack_size, true); duke@435: } duke@435: duke@435: // accessors duke@435: bool is_advanced_heap_walk() const { return _is_advanced_heap_walk; } duke@435: JvmtiTagMap* tag_map() const { return _tag_map; } duke@435: Handle initial_object() const { return _initial_object; } duke@435: duke@435: bool is_following_references() const { return _following_object_refs; } duke@435: duke@435: bool is_reporting_primitive_fields() const { return _reporting_primitive_fields; } duke@435: bool is_reporting_primitive_array_values() const { return _reporting_primitive_array_values; } duke@435: bool is_reporting_string_values() const { return _reporting_string_values; } duke@435: duke@435: GrowableArray* visit_stack() const { return _visit_stack; } duke@435: duke@435: // iterate over the various object types duke@435: inline bool iterate_over_array(oop o); duke@435: inline bool iterate_over_type_array(oop o); duke@435: inline bool iterate_over_class(klassOop o); duke@435: inline bool iterate_over_object(oop o); duke@435: duke@435: // root collection duke@435: inline bool collect_simple_roots(); duke@435: inline bool collect_stack_roots(); duke@435: inline bool collect_stack_roots(JavaThread* java_thread, JNILocalRootsClosure* blk); duke@435: duke@435: // visit an object duke@435: inline bool visit(oop o); duke@435: duke@435: public: duke@435: VM_HeapWalkOperation(JvmtiTagMap* tag_map, duke@435: Handle initial_object, duke@435: BasicHeapWalkContext callbacks, duke@435: const void* user_data); duke@435: duke@435: VM_HeapWalkOperation(JvmtiTagMap* tag_map, duke@435: Handle initial_object, duke@435: AdvancedHeapWalkContext callbacks, duke@435: const void* user_data); duke@435: duke@435: ~VM_HeapWalkOperation(); duke@435: duke@435: VMOp_Type type() const { return VMOp_HeapWalkOperation; } duke@435: void doit(); duke@435: }; duke@435: duke@435: duke@435: VM_HeapWalkOperation::VM_HeapWalkOperation(JvmtiTagMap* tag_map, duke@435: Handle initial_object, duke@435: BasicHeapWalkContext callbacks, duke@435: const void* user_data) { duke@435: _is_advanced_heap_walk = false; duke@435: _tag_map = tag_map; duke@435: _initial_object = initial_object; duke@435: _following_object_refs = (callbacks.object_ref_callback() != NULL); duke@435: _reporting_primitive_fields = false; duke@435: _reporting_primitive_array_values = false; duke@435: _reporting_string_values = false; duke@435: _visit_stack = create_visit_stack(); duke@435: duke@435: duke@435: CallbackInvoker::initialize_for_basic_heap_walk(tag_map, _visit_stack, user_data, callbacks); duke@435: } duke@435: duke@435: VM_HeapWalkOperation::VM_HeapWalkOperation(JvmtiTagMap* tag_map, duke@435: Handle initial_object, duke@435: AdvancedHeapWalkContext callbacks, duke@435: const void* user_data) { duke@435: _is_advanced_heap_walk = true; duke@435: _tag_map = tag_map; duke@435: _initial_object = initial_object; duke@435: _following_object_refs = true; duke@435: _reporting_primitive_fields = (callbacks.primitive_field_callback() != NULL);; duke@435: _reporting_primitive_array_values = (callbacks.array_primitive_value_callback() != NULL);; duke@435: _reporting_string_values = (callbacks.string_primitive_value_callback() != NULL);; duke@435: _visit_stack = create_visit_stack(); duke@435: duke@435: CallbackInvoker::initialize_for_advanced_heap_walk(tag_map, _visit_stack, user_data, callbacks); duke@435: } duke@435: duke@435: VM_HeapWalkOperation::~VM_HeapWalkOperation() { duke@435: if (_following_object_refs) { duke@435: assert(_visit_stack != NULL, "checking"); duke@435: delete _visit_stack; duke@435: _visit_stack = NULL; duke@435: } duke@435: } duke@435: duke@435: // an array references its class and has a reference to duke@435: // each element in the array duke@435: inline bool VM_HeapWalkOperation::iterate_over_array(oop o) { duke@435: objArrayOop array = objArrayOop(o); duke@435: if (array->klass() == Universe::systemObjArrayKlassObj()) { duke@435: // filtered out duke@435: return true; duke@435: } duke@435: duke@435: // array reference to its class duke@435: oop mirror = objArrayKlass::cast(array->klass())->java_mirror(); duke@435: if (!CallbackInvoker::report_class_reference(o, mirror)) { duke@435: return false; duke@435: } duke@435: duke@435: // iterate over the array and report each reference to a duke@435: // non-null element duke@435: for (int index=0; indexlength(); index++) { duke@435: oop elem = array->obj_at(index); duke@435: if (elem == NULL) { duke@435: continue; duke@435: } duke@435: duke@435: // report the array reference o[index] = elem duke@435: if (!CallbackInvoker::report_array_element_reference(o, elem, index)) { duke@435: return false; duke@435: } duke@435: } duke@435: return true; duke@435: } duke@435: duke@435: // a type array references its class duke@435: inline bool VM_HeapWalkOperation::iterate_over_type_array(oop o) { duke@435: klassOop k = o->klass(); duke@435: oop mirror = Klass::cast(k)->java_mirror(); duke@435: if (!CallbackInvoker::report_class_reference(o, mirror)) { duke@435: return false; duke@435: } duke@435: duke@435: // report the array contents if required duke@435: if (is_reporting_primitive_array_values()) { duke@435: if (!CallbackInvoker::report_primitive_array_values(o)) { duke@435: return false; duke@435: } duke@435: } duke@435: return true; duke@435: } duke@435: duke@435: // verify that a static oop field is in range coleenp@548: static inline bool verify_static_oop(instanceKlass* ik, coleenp@548: klassOop k, int offset) { coleenp@548: address obj_p = (address)k + offset; coleenp@548: address start = (address)ik->start_of_static_fields(); coleenp@548: address end = start + (ik->static_oop_field_size() * heapOopSize); duke@435: assert(end >= start, "sanity check"); duke@435: duke@435: if (obj_p >= start && obj_p < end) { duke@435: return true; duke@435: } else { duke@435: return false; duke@435: } duke@435: } duke@435: duke@435: // a class references its super class, interfaces, class loader, ... duke@435: // and finally its static fields duke@435: inline bool VM_HeapWalkOperation::iterate_over_class(klassOop k) { duke@435: int i; duke@435: Klass* klass = klassOop(k)->klass_part(); duke@435: duke@435: if (klass->oop_is_instance()) { duke@435: instanceKlass* ik = instanceKlass::cast(k); duke@435: duke@435: // ignore the class if it's has been initialized yet duke@435: if (!ik->is_linked()) { duke@435: return true; duke@435: } duke@435: duke@435: // get the java mirror duke@435: oop mirror = klass->java_mirror(); duke@435: duke@435: // super (only if something more interesting than java.lang.Object) duke@435: klassOop java_super = ik->java_super(); duke@435: if (java_super != NULL && java_super != SystemDictionary::object_klass()) { duke@435: oop super = Klass::cast(java_super)->java_mirror(); duke@435: if (!CallbackInvoker::report_superclass_reference(mirror, super)) { duke@435: return false; duke@435: } duke@435: } duke@435: duke@435: // class loader duke@435: oop cl = ik->class_loader(); duke@435: if (cl != NULL) { duke@435: if (!CallbackInvoker::report_class_loader_reference(mirror, cl)) { duke@435: return false; duke@435: } duke@435: } duke@435: duke@435: // protection domain duke@435: oop pd = ik->protection_domain(); duke@435: if (pd != NULL) { duke@435: if (!CallbackInvoker::report_protection_domain_reference(mirror, pd)) { duke@435: return false; duke@435: } duke@435: } duke@435: duke@435: // signers duke@435: oop signers = ik->signers(); duke@435: if (signers != NULL) { duke@435: if (!CallbackInvoker::report_signers_reference(mirror, signers)) { duke@435: return false; duke@435: } duke@435: } duke@435: duke@435: // references from the constant pool duke@435: { duke@435: const constantPoolOop pool = ik->constants(); duke@435: for (int i = 1; i < pool->length(); i++) { duke@435: constantTag tag = pool->tag_at(i).value(); duke@435: if (tag.is_string() || tag.is_klass()) { duke@435: oop entry; duke@435: if (tag.is_string()) { duke@435: entry = pool->resolved_string_at(i); duke@435: assert(java_lang_String::is_instance(entry), "must be string"); duke@435: } else { duke@435: entry = Klass::cast(pool->resolved_klass_at(i))->java_mirror(); duke@435: } duke@435: if (!CallbackInvoker::report_constant_pool_reference(mirror, entry, (jint)i)) { duke@435: return false; duke@435: } duke@435: } duke@435: } duke@435: } duke@435: duke@435: // interfaces duke@435: // (These will already have been reported as references from the constant pool duke@435: // but are specified by IterateOverReachableObjects and must be reported). duke@435: objArrayOop interfaces = ik->local_interfaces(); duke@435: for (i = 0; i < interfaces->length(); i++) { duke@435: oop interf = Klass::cast((klassOop)interfaces->obj_at(i))->java_mirror(); duke@435: if (interf == NULL) { duke@435: continue; duke@435: } duke@435: if (!CallbackInvoker::report_interface_reference(mirror, interf)) { duke@435: return false; duke@435: } duke@435: } duke@435: duke@435: // iterate over the static fields duke@435: duke@435: ClassFieldMap* field_map = ClassFieldMap::create_map_of_static_fields(k); duke@435: for (i=0; ifield_count(); i++) { duke@435: ClassFieldDescriptor* field = field_map->field_at(i); duke@435: char type = field->field_type(); duke@435: if (!is_primitive_field_type(type)) { coleenp@548: oop fld_o = k->obj_field(field->field_offset()); coleenp@548: assert(verify_static_oop(ik, k, field->field_offset()), "sanity check"); duke@435: if (fld_o != NULL) { duke@435: int slot = field->field_index(); duke@435: if (!CallbackInvoker::report_static_field_reference(mirror, fld_o, slot)) { duke@435: delete field_map; duke@435: return false; duke@435: } duke@435: } duke@435: } else { duke@435: if (is_reporting_primitive_fields()) { duke@435: address addr = (address)k + field->field_offset(); duke@435: int slot = field->field_index(); duke@435: if (!CallbackInvoker::report_primitive_static_field(mirror, slot, addr, type)) { duke@435: delete field_map; duke@435: return false; duke@435: } duke@435: } duke@435: } duke@435: } duke@435: delete field_map; duke@435: duke@435: return true; duke@435: } duke@435: duke@435: return true; duke@435: } duke@435: duke@435: // an object references a class and its instance fields duke@435: // (static fields are ignored here as we report these as duke@435: // references from the class). duke@435: inline bool VM_HeapWalkOperation::iterate_over_object(oop o) { duke@435: // reference to the class duke@435: if (!CallbackInvoker::report_class_reference(o, Klass::cast(o->klass())->java_mirror())) { duke@435: return false; duke@435: } duke@435: duke@435: // iterate over instance fields duke@435: ClassFieldMap* field_map = JvmtiCachedClassFieldMap::get_map_of_instance_fields(o); duke@435: for (int i=0; ifield_count(); i++) { duke@435: ClassFieldDescriptor* field = field_map->field_at(i); duke@435: char type = field->field_type(); duke@435: if (!is_primitive_field_type(type)) { coleenp@548: oop fld_o = o->obj_field(field->field_offset()); duke@435: if (fld_o != NULL) { duke@435: // reflection code may have a reference to a klassOop. duke@435: // - see sun.reflect.UnsafeStaticFieldAccessorImpl and sun.misc.Unsafe duke@435: if (fld_o->is_klass()) { duke@435: klassOop k = (klassOop)fld_o; duke@435: fld_o = Klass::cast(k)->java_mirror(); duke@435: } duke@435: int slot = field->field_index(); duke@435: if (!CallbackInvoker::report_field_reference(o, fld_o, slot)) { duke@435: return false; duke@435: } duke@435: } duke@435: } else { duke@435: if (is_reporting_primitive_fields()) { duke@435: // primitive instance field duke@435: address addr = (address)o + field->field_offset(); duke@435: int slot = field->field_index(); duke@435: if (!CallbackInvoker::report_primitive_instance_field(o, slot, addr, type)) { duke@435: return false; duke@435: } duke@435: } duke@435: } duke@435: } duke@435: duke@435: // if the object is a java.lang.String duke@435: if (is_reporting_string_values() && duke@435: o->klass() == SystemDictionary::string_klass()) { duke@435: if (!CallbackInvoker::report_string_value(o)) { duke@435: return false; duke@435: } duke@435: } duke@435: return true; duke@435: } duke@435: duke@435: duke@435: // collects all simple (non-stack) roots. duke@435: // if there's a heap root callback provided then the callback is duke@435: // invoked for each simple root. duke@435: // if an object reference callback is provided then all simple duke@435: // roots are pushed onto the marking stack so that they can be duke@435: // processed later duke@435: // duke@435: inline bool VM_HeapWalkOperation::collect_simple_roots() { duke@435: SimpleRootsClosure blk; duke@435: duke@435: // JNI globals duke@435: blk.set_kind(JVMTI_HEAP_REFERENCE_JNI_GLOBAL); duke@435: JNIHandles::oops_do(&blk); duke@435: if (blk.stopped()) { duke@435: return false; duke@435: } duke@435: duke@435: // Preloaded classes and loader from the system dictionary duke@435: blk.set_kind(JVMTI_HEAP_REFERENCE_SYSTEM_CLASS); duke@435: SystemDictionary::always_strong_oops_do(&blk); duke@435: if (blk.stopped()) { duke@435: return false; duke@435: } duke@435: duke@435: // Inflated monitors duke@435: blk.set_kind(JVMTI_HEAP_REFERENCE_MONITOR); duke@435: ObjectSynchronizer::oops_do(&blk); duke@435: if (blk.stopped()) { duke@435: return false; duke@435: } duke@435: duke@435: // Threads duke@435: for (JavaThread* thread = Threads::first(); thread != NULL ; thread = thread->next()) { duke@435: oop threadObj = thread->threadObj(); duke@435: if (threadObj != NULL && !thread->is_exiting() && !thread->is_hidden_from_external_view()) { duke@435: bool cont = CallbackInvoker::report_simple_root(JVMTI_HEAP_REFERENCE_THREAD, threadObj); duke@435: if (!cont) { duke@435: return false; duke@435: } duke@435: } duke@435: } duke@435: duke@435: // Other kinds of roots maintained by HotSpot duke@435: // Many of these won't be visible but others (such as instances of important duke@435: // exceptions) will be visible. duke@435: blk.set_kind(JVMTI_HEAP_REFERENCE_OTHER); duke@435: Universe::oops_do(&blk); duke@435: return true; duke@435: } duke@435: duke@435: // Walk the stack of a given thread and find all references (locals duke@435: // and JNI calls) and report these as stack references duke@435: inline bool VM_HeapWalkOperation::collect_stack_roots(JavaThread* java_thread, duke@435: JNILocalRootsClosure* blk) duke@435: { duke@435: oop threadObj = java_thread->threadObj(); duke@435: assert(threadObj != NULL, "sanity check"); duke@435: duke@435: // only need to get the thread's tag once per thread duke@435: jlong thread_tag = tag_for(_tag_map, threadObj); duke@435: duke@435: // also need the thread id duke@435: jlong tid = java_lang_Thread::thread_id(threadObj); duke@435: duke@435: duke@435: if (java_thread->has_last_Java_frame()) { duke@435: duke@435: // vframes are resource allocated duke@435: Thread* current_thread = Thread::current(); duke@435: ResourceMark rm(current_thread); duke@435: HandleMark hm(current_thread); duke@435: duke@435: RegisterMap reg_map(java_thread); duke@435: frame f = java_thread->last_frame(); duke@435: vframe* vf = vframe::new_vframe(&f, ®_map, java_thread); duke@435: duke@435: bool is_top_frame = true; duke@435: int depth = 0; duke@435: frame* last_entry_frame = NULL; duke@435: duke@435: while (vf != NULL) { duke@435: if (vf->is_java_frame()) { duke@435: duke@435: // java frame (interpreted, compiled, ...) duke@435: javaVFrame *jvf = javaVFrame::cast(vf); duke@435: duke@435: // the jmethodID duke@435: jmethodID method = jvf->method()->jmethod_id(); duke@435: duke@435: if (!(jvf->method()->is_native())) { duke@435: jlocation bci = (jlocation)jvf->bci(); duke@435: StackValueCollection* locals = jvf->locals(); duke@435: for (int slot=0; slotsize(); slot++) { duke@435: if (locals->at(slot)->type() == T_OBJECT) { duke@435: oop o = locals->obj_at(slot)(); duke@435: if (o == NULL) { duke@435: continue; duke@435: } duke@435: duke@435: // stack reference duke@435: if (!CallbackInvoker::report_stack_ref_root(thread_tag, tid, depth, method, duke@435: bci, slot, o)) { duke@435: return false; duke@435: } duke@435: } duke@435: } duke@435: } else { duke@435: blk->set_context(thread_tag, tid, depth, method); duke@435: if (is_top_frame) { duke@435: // JNI locals for the top frame. duke@435: java_thread->active_handles()->oops_do(blk); duke@435: } else { duke@435: if (last_entry_frame != NULL) { duke@435: // JNI locals for the entry frame duke@435: assert(last_entry_frame->is_entry_frame(), "checking"); duke@435: last_entry_frame->entry_frame_call_wrapper()->handles()->oops_do(blk); duke@435: } duke@435: } duke@435: } duke@435: last_entry_frame = NULL; duke@435: depth++; duke@435: } else { duke@435: // externalVFrame - for an entry frame then we report the JNI locals duke@435: // when we find the corresponding javaVFrame duke@435: frame* fr = vf->frame_pointer(); duke@435: assert(fr != NULL, "sanity check"); duke@435: if (fr->is_entry_frame()) { duke@435: last_entry_frame = fr; duke@435: } duke@435: } duke@435: duke@435: vf = vf->sender(); duke@435: is_top_frame = false; duke@435: } duke@435: } else { duke@435: // no last java frame but there may be JNI locals duke@435: blk->set_context(thread_tag, tid, 0, (jmethodID)NULL); duke@435: java_thread->active_handles()->oops_do(blk); duke@435: } duke@435: return true; duke@435: } duke@435: duke@435: duke@435: // collects all stack roots - for each thread it walks the execution duke@435: // stack to find all references and local JNI refs. duke@435: inline bool VM_HeapWalkOperation::collect_stack_roots() { duke@435: JNILocalRootsClosure blk; duke@435: for (JavaThread* thread = Threads::first(); thread != NULL ; thread = thread->next()) { duke@435: oop threadObj = thread->threadObj(); duke@435: if (threadObj != NULL && !thread->is_exiting() && !thread->is_hidden_from_external_view()) { duke@435: if (!collect_stack_roots(thread, &blk)) { duke@435: return false; duke@435: } duke@435: } duke@435: } duke@435: return true; duke@435: } duke@435: duke@435: // visit an object duke@435: // first mark the object as visited duke@435: // second get all the outbound references from this object (in other words, all duke@435: // the objects referenced by this object). duke@435: // duke@435: bool VM_HeapWalkOperation::visit(oop o) { duke@435: // mark object as visited duke@435: assert(!ObjectMarker::visited(o), "can't visit same object more than once"); duke@435: ObjectMarker::mark(o); duke@435: duke@435: // instance duke@435: if (o->is_instance()) { duke@435: if (o->klass() == SystemDictionary::class_klass()) { duke@435: o = klassOop_if_java_lang_Class(o); duke@435: if (o->is_klass()) { duke@435: // a java.lang.Class duke@435: return iterate_over_class(klassOop(o)); duke@435: } duke@435: } else { duke@435: return iterate_over_object(o); duke@435: } duke@435: } duke@435: duke@435: // object array duke@435: if (o->is_objArray()) { duke@435: return iterate_over_array(o); duke@435: } duke@435: duke@435: // type array duke@435: if (o->is_typeArray()) { duke@435: return iterate_over_type_array(o); duke@435: } duke@435: duke@435: return true; duke@435: } duke@435: duke@435: void VM_HeapWalkOperation::doit() { duke@435: ResourceMark rm; duke@435: ObjectMarkerController marker; duke@435: ClassFieldMapCacheMark cm; duke@435: duke@435: assert(visit_stack()->is_empty(), "visit stack must be empty"); duke@435: duke@435: // the heap walk starts with an initial object or the heap roots duke@435: if (initial_object().is_null()) { duke@435: if (!collect_simple_roots()) return; duke@435: if (!collect_stack_roots()) return; duke@435: } else { duke@435: visit_stack()->push(initial_object()()); duke@435: } duke@435: duke@435: // object references required duke@435: if (is_following_references()) { duke@435: duke@435: // visit each object until all reachable objects have been duke@435: // visited or the callback asked to terminate the iteration. duke@435: while (!visit_stack()->is_empty()) { duke@435: oop o = visit_stack()->pop(); duke@435: if (!ObjectMarker::visited(o)) { duke@435: if (!visit(o)) { duke@435: break; duke@435: } duke@435: } duke@435: } duke@435: } duke@435: } duke@435: duke@435: // iterate over all objects that are reachable from a set of roots duke@435: void JvmtiTagMap::iterate_over_reachable_objects(jvmtiHeapRootCallback heap_root_callback, duke@435: jvmtiStackReferenceCallback stack_ref_callback, duke@435: jvmtiObjectReferenceCallback object_ref_callback, duke@435: const void* user_data) { duke@435: MutexLocker ml(Heap_lock); duke@435: BasicHeapWalkContext context(heap_root_callback, stack_ref_callback, object_ref_callback); duke@435: VM_HeapWalkOperation op(this, Handle(), context, user_data); duke@435: VMThread::execute(&op); duke@435: } duke@435: duke@435: // iterate over all objects that are reachable from a given object duke@435: void JvmtiTagMap::iterate_over_objects_reachable_from_object(jobject object, duke@435: jvmtiObjectReferenceCallback object_ref_callback, duke@435: const void* user_data) { duke@435: oop obj = JNIHandles::resolve(object); duke@435: Handle initial_object(Thread::current(), obj); duke@435: duke@435: MutexLocker ml(Heap_lock); duke@435: BasicHeapWalkContext context(NULL, NULL, object_ref_callback); duke@435: VM_HeapWalkOperation op(this, initial_object, context, user_data); duke@435: VMThread::execute(&op); duke@435: } duke@435: duke@435: // follow references from an initial object or the GC roots duke@435: void JvmtiTagMap::follow_references(jint heap_filter, duke@435: KlassHandle klass, duke@435: jobject object, duke@435: const jvmtiHeapCallbacks* callbacks, duke@435: const void* user_data) duke@435: { duke@435: oop obj = JNIHandles::resolve(object); duke@435: Handle initial_object(Thread::current(), obj); duke@435: duke@435: MutexLocker ml(Heap_lock); duke@435: AdvancedHeapWalkContext context(heap_filter, klass, callbacks); duke@435: VM_HeapWalkOperation op(this, initial_object, context, user_data); duke@435: VMThread::execute(&op); duke@435: } duke@435: duke@435: duke@435: // called post-GC duke@435: // - for each JVMTI environment with an object tag map, call its rehash duke@435: // function to re-sync with the new object locations. duke@435: void JvmtiTagMap::gc_epilogue(bool full) { duke@435: assert(SafepointSynchronize::is_at_safepoint(), "must be executed at a safepoint"); duke@435: if (JvmtiEnv::environments_might_exist()) { duke@435: // re-obtain the memory region for the young generation (might duke@435: // changed due to adaptive resizing policy) duke@435: get_young_generation(); duke@435: duke@435: JvmtiEnvIterator it; duke@435: for (JvmtiEnvBase* env = it.first(); env != NULL; env = it.next(env)) { duke@435: JvmtiTagMap* tag_map = env->tag_map(); duke@435: if (tag_map != NULL && !tag_map->is_empty()) { duke@435: TraceTime t(full ? "JVMTI Full Rehash " : "JVMTI Rehash ", TraceJVMTIObjectTagging); duke@435: if (full) { duke@435: tag_map->rehash(0, n_hashmaps); duke@435: } else { duke@435: tag_map->rehash(0, 0); // tag map for young gen only duke@435: } duke@435: } duke@435: } duke@435: } duke@435: } duke@435: duke@435: // CMS has completed referencing processing so we may have JNI weak refs duke@435: // to objects in the CMS generation that have been GC'ed. duke@435: void JvmtiTagMap::cms_ref_processing_epilogue() { duke@435: assert(SafepointSynchronize::is_at_safepoint(), "must be executed at a safepoint"); duke@435: assert(UseConcMarkSweepGC, "should only be used with CMS"); duke@435: if (JvmtiEnv::environments_might_exist()) { duke@435: JvmtiEnvIterator it; duke@435: for (JvmtiEnvBase* env = it.first(); env != NULL; env = it.next(env)) { duke@435: JvmtiTagMap* tag_map = ((JvmtiEnvBase *)env)->tag_map(); duke@435: if (tag_map != NULL && !tag_map->is_empty()) { duke@435: TraceTime t("JVMTI Rehash (CMS) ", TraceJVMTIObjectTagging); duke@435: tag_map->rehash(1, n_hashmaps); // assume CMS not used in young gen duke@435: } duke@435: } duke@435: } duke@435: } duke@435: duke@435: duke@435: // For each entry in the hashmaps 'start' to 'end' : duke@435: // duke@435: // 1. resolve the JNI weak reference duke@435: // duke@435: // 2. If it resolves to NULL it means the object has been freed so the entry duke@435: // is removed, the weak reference destroyed, and the object free event is duke@435: // posted (if enabled). duke@435: // duke@435: // 3. If the weak reference resolves to an object then we re-hash the object duke@435: // to see if it has moved or has been promoted (from the young to the old duke@435: // generation for example). duke@435: // duke@435: void JvmtiTagMap::rehash(int start, int end) { duke@435: duke@435: // does this environment have the OBJECT_FREE event enabled duke@435: bool post_object_free = env()->is_enabled(JVMTI_EVENT_OBJECT_FREE); duke@435: duke@435: // counters used for trace message duke@435: int freed = 0; duke@435: int moved = 0; duke@435: int promoted = 0; duke@435: duke@435: // we assume there are two hashmaps - one for the young generation duke@435: // and the other for all other spaces. duke@435: assert(n_hashmaps == 2, "not implemented"); duke@435: JvmtiTagHashmap* young_hashmap = _hashmap[0]; duke@435: JvmtiTagHashmap* other_hashmap = _hashmap[1]; duke@435: duke@435: // reenable sizing (if disabled) duke@435: young_hashmap->set_resizing_enabled(true); duke@435: other_hashmap->set_resizing_enabled(true); duke@435: duke@435: // when re-hashing the hashmap corresponding to the young generation we duke@435: // collect the entries corresponding to objects that have been promoted. duke@435: JvmtiTagHashmapEntry* promoted_entries = NULL; duke@435: duke@435: if (end >= n_hashmaps) { duke@435: end = n_hashmaps - 1; duke@435: } duke@435: duke@435: for (int i=start; i <= end; i++) { duke@435: JvmtiTagHashmap* hashmap = _hashmap[i]; duke@435: duke@435: // if the hashmap is empty then we can skip it duke@435: if (hashmap->_entry_count == 0) { duke@435: continue; duke@435: } duke@435: duke@435: // now iterate through each entry in the table duke@435: duke@435: JvmtiTagHashmapEntry** table = hashmap->table(); duke@435: int size = hashmap->size(); duke@435: duke@435: for (int pos=0; posnext(); duke@435: duke@435: jweak ref = entry->object(); duke@435: oop oop = JNIHandles::resolve(ref); duke@435: duke@435: // has object been GC'ed duke@435: if (oop == NULL) { duke@435: // grab the tag duke@435: jlong tag = entry->tag(); duke@435: guarantee(tag != 0, "checking"); duke@435: duke@435: // remove GC'ed entry from hashmap and return the duke@435: // entry to the free list duke@435: hashmap->remove(prev, pos, entry); duke@435: destroy_entry(entry); duke@435: duke@435: // destroy the weak ref duke@435: JNIHandles::destroy_weak_global(ref); duke@435: duke@435: // post the event to the profiler duke@435: if (post_object_free) { duke@435: JvmtiExport::post_object_free(env(), tag); duke@435: } duke@435: duke@435: freed++; duke@435: entry = next; duke@435: continue; duke@435: } duke@435: duke@435: // if this is the young hashmap then the object is either promoted duke@435: // or moved. duke@435: // if this is the other hashmap then the object is moved. duke@435: duke@435: bool same_gen; duke@435: if (i == 0) { duke@435: assert(hashmap == young_hashmap, "checking"); duke@435: same_gen = is_in_young(oop); duke@435: } else { duke@435: same_gen = true; duke@435: } duke@435: duke@435: duke@435: if (same_gen) { duke@435: // if the object has moved then re-hash it and move its duke@435: // entry to its new location. duke@435: unsigned int new_pos = JvmtiTagHashmap::hash(oop, size); duke@435: if (new_pos != (unsigned int)pos) { duke@435: if (prev == NULL) { duke@435: table[pos] = next; duke@435: } else { duke@435: prev->set_next(next); duke@435: } duke@435: entry->set_next(table[new_pos]); duke@435: table[new_pos] = entry; duke@435: moved++; duke@435: } else { duke@435: // object didn't move duke@435: prev = entry; duke@435: } duke@435: } else { duke@435: // object has been promoted so remove the entry from the duke@435: // young hashmap duke@435: assert(hashmap == young_hashmap, "checking"); duke@435: hashmap->remove(prev, pos, entry); duke@435: duke@435: // move the entry to the promoted list duke@435: entry->set_next(promoted_entries); duke@435: promoted_entries = entry; duke@435: } duke@435: duke@435: entry = next; duke@435: } duke@435: } duke@435: } duke@435: duke@435: duke@435: // add the entries, corresponding to the promoted objects, to the duke@435: // other hashmap. duke@435: JvmtiTagHashmapEntry* entry = promoted_entries; duke@435: while (entry != NULL) { duke@435: oop o = JNIHandles::resolve(entry->object()); duke@435: assert(hashmap_for(o) == other_hashmap, "checking"); duke@435: JvmtiTagHashmapEntry* next = entry->next(); duke@435: other_hashmap->add(o, entry); duke@435: entry = next; duke@435: promoted++; duke@435: } duke@435: duke@435: // stats duke@435: if (TraceJVMTIObjectTagging) { duke@435: int total_moves = promoted + moved; duke@435: duke@435: int post_total = 0; duke@435: for (int i=0; i_entry_count; duke@435: } duke@435: int pre_total = post_total + freed; duke@435: duke@435: tty->print("(%d->%d, %d freed, %d promoted, %d total moves)", duke@435: pre_total, post_total, freed, promoted, total_moves); duke@435: } duke@435: }