/openjdk7/jdk/src/share/classes/javax/management/modelmbean/ |
H A D | ModelMBean.java | 105 * @param mr Object that is the managed resource 119 public void setManagedResource(Object mr, String mr_type) argument
|
/openjdk7/hotspot/src/share/vm/gc_implementation/shared/ |
H A D | immutableSpace.cpp | 32 void ImmutableSpace::initialize(MemRegion mr) { argument 33 HeapWord* bottom = mr.start(); 34 HeapWord* end = mr.end();
|
H A D | spaceDecorator.cpp | 82 // Simply mangle the MemRegion mr. 83 void SpaceMangler::mangle_region(MemRegion mr) { argument 87 gclog_or_tty->print("Mangling [0x%x to 0x%x)", mr.start(), mr.end()); 89 Copy::fill_to_words(mr.start(), mr.word_size(), badHeapWord);
|
H A D | mutableSpace.hpp | 58 void numa_setup_pages(MemRegion mr, bool clear_space); 59 void pretouch_pages(MemRegion mr); 61 void set_last_setup_region(MemRegion mr) { _last_setup_region = mr; } argument 87 virtual void initialize(MemRegion mr, 95 void initialize(MemRegion mr, bool clear_space, bool reset_top); 112 virtual void mangle_region(MemRegion mr) PRODUCT_RETURN;
|
H A D | parGCAllocBuffer.cpp | 164 void ParGCAllocBufferWithBOT::fill_region_with_block(MemRegion mr, argument 166 CollectedHeap::fill_with_object(mr); 168 _bt.alloc_block(mr.start(), mr.end()); 170 _bt.BlockOffsetArray::alloc_block(mr.start(), mr.end());
|
H A D | mutableSpace.cpp | 45 void MutableSpace::numa_setup_pages(MemRegion mr, bool clear_space) { argument 46 if (!mr.is_empty()) { 48 HeapWord *start = (HeapWord*)round_to((intptr_t) mr.start(), page_size); 49 HeapWord *end = (HeapWord*)round_down((intptr_t) mr.end(), page_size); 61 void MutableSpace::pretouch_pages(MemRegion mr) { argument 62 for (volatile char *p = (char*)mr.start(); p < (char*)mr.end(); p += os::vm_page_size()) { 67 void MutableSpace::initialize(MemRegion mr, argument 72 assert(Universe::on_page_boundary(mr.start()) && Universe::on_page_boundary(mr 160 mangle_region(MemRegion mr) argument [all...] |
/openjdk7/hotspot/src/share/vm/memory/ |
H A D | barrierSet.inline.hpp | 79 void BarrierSet::write_region(MemRegion mr) { argument 81 ((CardTableModRefBS*)this)->inline_write_region(mr); 83 write_region_work(mr);
|
H A D | memRegion.hpp | 54 MemRegion(const MemRegion& mr): _start(mr._start), _word_size(mr._word_size) {} argument 95 virtual void do_MemRegion(MemRegion mr) = 0;
|
H A D | modRefBarrierSet.hpp | 78 void read_ref_array(MemRegion mr) { argument 81 void read_prim_array(MemRegion mr) { argument 84 void write_prim_array(MemRegion mr) { argument 87 void read_region(MemRegion mr) { argument 91 // Causes all refs in "mr" to be assumed to be modified. If "whole_heap" 94 virtual void invalidate(MemRegion mr, bool whole_heap = false) = 0; 96 // The caller guarantees that "mr" contains no references. (Perhaps it's 98 virtual void clear(MemRegion mr) = 0;
|
H A D | cardTableRS.hpp | 141 void verify_aligned_region_empty(MemRegion mr); 143 void clear(MemRegion mr) { _ct_bs->clear(mr); } argument 146 void invalidate(MemRegion mr, bool whole_heap = false) { argument 147 _ct_bs->invalidate(mr, whole_heap); 188 void do_MemRegion(MemRegion mr);
|
/openjdk7/hotspot/src/share/vm/oops/ |
H A D | instanceRefKlass.hpp | 71 int oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) { argument 72 return oop_oop_iterate_v_m(obj, blk, mr); 77 int oop_oop_iterate##nv_suffix##_m(oop obj, OopClosureType* blk, MemRegion mr);
|
H A D | instanceMirrorKlass.hpp | 92 int oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) { argument 93 return oop_oop_iterate_v_m(obj, blk, mr); 98 int oop_oop_iterate##nv_suffix##_m(oop obj, OopClosureType* blk, MemRegion mr);
|
H A D | arrayKlassKlass.cpp | 111 int arrayKlassKlass::oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) { argument 115 if (mr.contains(addr)) blk->do_oop(addr); 117 if (mr.contains(addr)) blk->do_oop(addr); 119 if (mr.contains(addr)) blk->do_oop(addr); 120 ak->vtable()->oop_oop_iterate_m(blk, mr); 121 return klassKlass::oop_oop_iterate_m(obj, blk, mr);
|
H A D | compiledICHolderKlass.cpp | 103 MemRegion mr) { 110 obj->oop_iterate_header(blk, mr); 114 if (mr.contains(adr)) blk->do_oop(adr); 116 if (mr.contains(adr)) blk->do_oop(adr); 102 oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) argument
|
H A D | klassKlass.cpp | 130 int klassKlass::oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) { argument 136 if (mr.contains(adr)) blk->do_oop(adr); 139 if (mr.contains(adr)) blk->do_oop(adr); 142 if (mr.contains(adr)) blk->do_oop(adr); 144 if (mr.contains(adr)) blk->do_oop(adr); 146 if (mr.contains(adr)) blk->do_oop(adr); 153 && (mr.contains(k->adr_subklass()) 154 || mr.contains(k->adr_next_sibling()))) { 157 obj->oop_iterate_header(blk, mr);
|
H A D | objArrayKlass.hpp | 118 int oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) { argument 119 return oop_oop_iterate_v_m(obj, blk, mr); 124 MemRegion mr); \
|
H A D | cpCacheKlass.cpp | 137 int constantPoolCacheKlass::oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) { argument 147 if (mr.contains(addr)) blk->do_oop(addr); 149 for (int i = 0; i < cache->length(); i++) cache->entry_at(i)->oop_iterate_m(blk, mr);
|
H A D | methodDataKlass.cpp | 131 int methodDataKlass::oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) { argument 138 obj->oop_iterate_header(blk, mr); 140 if (mr.contains(adr)) { 147 data->oop_iterate_m(blk, mr);
|
H A D | methodKlass.cpp | 162 int methodKlass::oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) { argument 172 if (mr.contains(adr)) blk->do_oop(adr); 175 if (mr.contains(adr)) blk->do_oop(adr);
|
H A D | constMethodKlass.cpp | 133 int constMethodKlass::oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) { argument 138 if (mr.contains(adr)) blk->do_oop(adr); 140 if (mr.contains(adr)) blk->do_oop(adr);
|
/openjdk7/hotspot/src/share/vm/gc_implementation/parallelScavenge/ |
H A D | objectStartArray.cpp | 79 void ObjectStartArray::set_covered_region(MemRegion mr) { argument 80 assert(_reserved_region.contains(mr), "MemRegion outside of reserved space"); 81 assert(_reserved_region.start() == mr.start(), "Attempt to move covered region"); 83 HeapWord* low_bound = mr.start(); 84 HeapWord* high_bound = mr.end(); 88 size_t requested_blocks_size_in_bytes = mr.word_size() / block_size_in_words; 94 _covered_region = mr;
|
/openjdk7/hotspot/src/share/vm/gc_implementation/g1/ |
H A D | g1SATBCardTableModRefBS.cpp | 122 G1SATBCardTableLoggingModRefBS::invalidate(MemRegion mr, bool whole_heap) { argument 123 jbyte* byte = byte_for(mr.start()); 124 jbyte* last_byte = byte_for(mr.last());
|
H A D | g1SATBCardTableModRefBS.hpp | 114 void invalidate(MemRegion mr, bool whole_heap = false); 116 void write_region_work(MemRegion mr) { invalidate(mr); } argument 117 void write_ref_array_work(MemRegion mr) { invalidate(mr); } argument
|
H A D | concurrentMark.inline.hpp | 80 inline void ConcurrentMark::count_region(MemRegion mr, HeapRegion* hr, argument 86 HeapWord* start = mr.start(); 87 HeapWord* end = mr.end(); 88 size_t region_size_bytes = mr.byte_size(); 93 assert(hr == g1h->heap_region_containing(mr.last()), "sanity"); 120 inline void ConcurrentMark::count_region(MemRegion mr, argument 125 count_region(mr, hr, marked_bytes_array, task_card_bm); 130 inline void ConcurrentMark::count_region(MemRegion mr, uint worker_id) { argument 131 HeapWord* addr = mr.start(); 133 count_region(mr, h 244 iterate(BitMapClosure* cl, MemRegion mr) argument [all...] |
/openjdk7/hotspot/src/share/vm/gc_implementation/parNew/ |
H A D | parCardTableModRefBS.cpp | 38 void CardTableModRefBS::non_clean_card_iterate_parallel_work(Space* sp, MemRegion mr, argument 68 process_stride(sp, mr, stride, n_strides, cl, ct, 75 intptr_t first_chunk_index = addr_to_chunk_index(mr.start()); 76 uintptr_t last_chunk_index = addr_to_chunk_index(mr.last());
|