/openjdk7/hotspot/src/share/vm/gc_implementation/shared/ |
H A D | mutableSpace.cpp | 45 void MutableSpace::numa_setup_pages(MemRegion mr, bool clear_space) { argument 46 if (!mr.is_empty()) { 48 HeapWord *start = (HeapWord*)round_to((intptr_t) mr.start(), page_size); 49 HeapWord *end = (HeapWord*)round_down((intptr_t) mr.end(), page_size); 61 void MutableSpace::pretouch_pages(MemRegion mr) { argument 62 for (volatile char *p = (char*)mr.start(); p < (char*)mr.end(); p += os::vm_page_size()) { 67 void MutableSpace::initialize(MemRegion mr, argument 72 assert(Universe::on_page_boundary(mr.start()) && Universe::on_page_boundary(mr 160 mangle_region(MemRegion mr) argument [all...] |
H A D | spaceDecorator.cpp | 82 // Simply mangle the MemRegion mr. 83 void SpaceMangler::mangle_region(MemRegion mr) { argument 87 gclog_or_tty->print("Mangling [0x%x to 0x%x)", mr.start(), mr.end()); 89 Copy::fill_to_words(mr.start(), mr.word_size(), badHeapWord);
|
H A D | mutableSpace.hpp | 58 void numa_setup_pages(MemRegion mr, bool clear_space); 59 void pretouch_pages(MemRegion mr); 61 void set_last_setup_region(MemRegion mr) { _last_setup_region = mr; } argument 87 virtual void initialize(MemRegion mr, 95 void initialize(MemRegion mr, bool clear_space, bool reset_top); 112 virtual void mangle_region(MemRegion mr) PRODUCT_RETURN;
|
H A D | immutableSpace.cpp | 32 void ImmutableSpace::initialize(MemRegion mr) { argument 33 HeapWord* bottom = mr.start(); 34 HeapWord* end = mr.end();
|
/openjdk7/hotspot/src/share/vm/memory/ |
H A D | modRefBarrierSet.hpp | 78 void read_ref_array(MemRegion mr) { argument 81 void read_prim_array(MemRegion mr) { argument 84 void write_prim_array(MemRegion mr) { argument 87 void read_region(MemRegion mr) { argument 91 // Causes all refs in "mr" to be assumed to be modified. If "whole_heap" 94 virtual void invalidate(MemRegion mr, bool whole_heap = false) = 0; 96 // The caller guarantees that "mr" contains no references. (Perhaps it's 98 virtual void clear(MemRegion mr) = 0;
|
H A D | barrierSet.hpp | 131 virtual void read_ref_array(MemRegion mr) = 0; 132 virtual void read_prim_array(MemRegion mr) = 0; 150 virtual void write_ref_array_work(MemRegion mr) = 0; 152 virtual void write_prim_array(MemRegion mr) = 0; 154 virtual void read_region(MemRegion mr) = 0; 159 inline void write_region(MemRegion mr); 161 virtual void write_region_work(MemRegion mr) = 0;
|
H A D | cardTableModRefBS.hpp | 147 // Returns the part of the region mr that doesn't intersect with 151 MemRegion committed_unique_to_self(int self, MemRegion mr) const; 173 // region mr in the given space and apply cl to any dirty sub-regions 174 // of mr. Dirty cards are _not_ cleared by the iterator method itself, 176 void non_clean_card_iterate_serial(MemRegion mr, MemRegionClosure* cl); 184 void non_clean_card_iterate_possibly_parallel(Space* sp, MemRegion mr, 190 void non_clean_card_iterate_parallel_work(Space* sp, MemRegion mr, 195 // Dirty the bytes corresponding to "mr" (not all of which must be 197 void dirty_MemRegion(MemRegion mr); 199 // Clear (to clean_card) the bytes entirely contained within "mr" (no 228 chunks_to_cover(MemRegion mr) argument 306 inline_write_region(MemRegion mr) argument 310 write_region_work(MemRegion mr) argument 315 inline_write_ref_array(MemRegion mr) argument 319 write_ref_array_work(MemRegion mr) argument 424 mod_card_iterate(MemRegion mr, MemRegionClosure* cl) argument [all...] |
H A D | cardTableModRefBS.cpp | 198 MemRegion mr) const { 199 MemRegion result = mr; 459 MemRegion mr, 462 if (!mr.is_empty()) { 500 non_clean_card_iterate_parallel_work(sp, mr, cl, ct, n_threads); 514 clear_cl.do_MemRegion(mr); 524 void CardTableModRefBS::non_clean_card_iterate_serial(MemRegion mr, argument 531 MemRegion mri = mr.intersection(_covered[i]); 559 void CardTableModRefBS::dirty_MemRegion(MemRegion mr) { argument 560 assert((HeapWord*)align_size_down((uintptr_t)mr 458 non_clean_card_iterate_possibly_parallel(Space* sp, MemRegion mr, OopsInGenClosure* cl, CardTableRS* ct) argument 570 invalidate(MemRegion mr, bool whole_heap) argument 579 clear_MemRegion(MemRegion mr) argument 593 clear(MemRegion mr) argument 600 dirty(MemRegion mr) argument 608 dirty_card_iterate(MemRegion mr, MemRegionClosure* cl) argument 633 dirty_card_range_after_reset(MemRegion mr, bool reset, int reset_val) argument 680 verify_region(MemRegion mr, jbyte val, bool val_equals) argument 705 verify_not_dirty_region(MemRegion mr) argument 709 verify_dirty_region(MemRegion mr) argument [all...] |
H A D | cardTableRS.hpp | 141 void verify_aligned_region_empty(MemRegion mr); 143 void clear(MemRegion mr) { _ct_bs->clear(mr); } argument 146 void invalidate(MemRegion mr, bool whole_heap = false) { argument 147 _ct_bs->invalidate(mr, whole_heap); 188 void do_MemRegion(MemRegion mr);
|
H A D | space.cpp | 75 void DirtyCardToOopClosure::walk_mem_region(MemRegion mr, argument 95 oop(bottom)->oop_iterate(_cl, mr); 100 // We get called with "mr" representing the dirty region 102 // we may need to extend the incoming "mr" to the right, 108 void DirtyCardToOopClosure::do_MemRegion(MemRegion mr) { argument 118 pCl->do_MemRegion(mr); 121 HeapWord* bottom = mr.start(); 122 HeapWord* last = mr.last(); 123 HeapWord* top = mr.end(); 135 NOT_PRODUCT(_last_bottom = mr 209 walk_mem_region(MemRegion mr, HeapWord* bottom, HeapWord* top) argument 357 mangle_region(MemRegion mr) argument 368 initialize(MemRegion mr, bool clear_space, bool mangle_space) argument 561 object_iterate_careful_m(MemRegion mr, ObjectClosureCareful* cl) argument 568 object_iterate_mem(MemRegion mr, UpwardsObjectClosure* cl) argument 628 object_iterate_mem(MemRegion mr, UpwardsObjectClosure* cl) argument 692 oop_iterate(MemRegion mr, OopClosure* blk) argument 969 OffsetTableContigSpace(BlockOffsetSharedArray* sharedOffsetArray, MemRegion mr) argument [all...] |
H A D | memRegion.hpp | 54 MemRegion(const MemRegion& mr): _start(mr._start), _word_size(mr._word_size) {} argument 95 virtual void do_MemRegion(MemRegion mr) = 0;
|
H A D | barrierSet.inline.hpp | 79 void BarrierSet::write_region(MemRegion mr) { argument 81 ((CardTableModRefBS*)this)->inline_write_region(mr); 83 write_region_work(mr);
|
H A D | genRemSet.hpp | 103 // the heap interval denoted by mr. If there are any 109 // [ region mr ) 111 virtual void verify_aligned_region_empty(MemRegion mr) = 0; 118 virtual void clear(MemRegion mr) = 0; 126 // Informs the RS that refs in the given "mr" may have changed 131 virtual void invalidate(MemRegion mr, bool whole_heap = false) = 0;
|
H A D | cardTableRS.cpp | 180 void ClearNoncleanCardWrapper::do_MemRegion(MemRegion mr) { argument 181 assert(mr.word_size() > 0, "Error"); 182 assert(_ct->is_aligned(mr.start()), "mr.start() should be card aligned"); 183 // mr.end() may not necessarily be card aligned. 184 jbyte* cur_entry = _ct->byte_for(mr.last()); 185 const jbyte* limit = _ct->byte_for(mr.start()); 186 HeapWord* end_of_non_clean = mr.end(); 220 // in the loop and, when we hit the left end of "mr", 222 // for "mr" 648 verify_aligned_region_empty(MemRegion mr) argument [all...] |
/openjdk7/hotspot/src/share/vm/gc_implementation/concurrentMarkSweep/ |
H A D | concurrentMarkSweepGeneration.inline.hpp | 76 inline void CMSBitMap::mark_range(MemRegion mr) { argument 77 NOT_PRODUCT(region_invariant(mr)); 79 _bm.set_range(heapWordToOffset(mr.start()), heapWordToOffset(mr.end()), 83 inline void CMSBitMap::clear_range(MemRegion mr) { argument 84 NOT_PRODUCT(region_invariant(mr)); 86 _bm.clear_range(heapWordToOffset(mr.start()), heapWordToOffset(mr.end()), 90 inline void CMSBitMap::par_mark_range(MemRegion mr) { argument 91 NOT_PRODUCT(region_invariant(mr)); 97 par_clear_range(MemRegion mr) argument 104 mark_large_range(MemRegion mr) argument 111 clear_large_range(MemRegion mr) argument 118 par_mark_large_range(MemRegion mr) argument 125 par_clear_large_range(MemRegion mr) argument 501 do_MemRegion(MemRegion mr) argument 510 do_MemRegion(MemRegion mr) argument [all...] |
/openjdk7/hotspot/src/share/vm/gc_implementation/parallelScavenge/ |
H A D | objectStartArray.cpp | 79 void ObjectStartArray::set_covered_region(MemRegion mr) { argument 80 assert(_reserved_region.contains(mr), "MemRegion outside of reserved space"); 81 assert(_reserved_region.start() == mr.start(), "Attempt to move covered region"); 83 HeapWord* low_bound = mr.start(); 84 HeapWord* high_bound = mr.end(); 88 size_t requested_blocks_size_in_bytes = mr.word_size() / block_size_in_words; 94 _covered_region = mr;
|
/openjdk7/hotspot/src/share/vm/gc_implementation/g1/ |
H A D | g1SATBCardTableModRefBS.hpp | 114 void invalidate(MemRegion mr, bool whole_heap = false); 116 void write_region_work(MemRegion mr) { invalidate(mr); } argument 117 void write_ref_array_work(MemRegion mr) { invalidate(mr); } argument
|
H A D | concurrentMark.inline.hpp | 80 inline void ConcurrentMark::count_region(MemRegion mr, HeapRegion* hr, argument 86 HeapWord* start = mr.start(); 87 HeapWord* end = mr.end(); 88 size_t region_size_bytes = mr.byte_size(); 93 assert(hr == g1h->heap_region_containing(mr.last()), "sanity"); 120 inline void ConcurrentMark::count_region(MemRegion mr, argument 125 count_region(mr, hr, marked_bytes_array, task_card_bm); 130 inline void ConcurrentMark::count_region(MemRegion mr, uint worker_id) { argument 131 HeapWord* addr = mr.start(); 133 count_region(mr, h 244 iterate(BitMapClosure* cl, MemRegion mr) argument [all...] |
H A D | heapRegion.cpp | 213 void HeapRegionDCTOC::walk_mem_region_with_cl(MemRegion mr, argument 236 oop_size = oop(bottom)->oop_iterate(cl2, mr); 268 oop(bottom)->oop_iterate(cl2, mr); 479 void HeapRegion::initialize(MemRegion mr, bool clear_space, bool mangle_space) { argument 480 G1OffsetTableContigSpace::initialize(mr, false, mangle_space); 490 MemRegion mr, bool is_zeroed) : 491 G1OffsetTableContigSpace(sharedOffsetArray, mr, is_zeroed), 507 _orig_end = mr.end(); 510 this->initialize(mr, !is_zeroed, SpaceDecorator::Mangle); 538 void HeapRegion::oops_in_mr_iterate(MemRegion mr, OopClosur argument 488 HeapRegion(uint hrs_index, G1BlockOffsetSharedArray* sharedOffsetArray, MemRegion mr, bool is_zeroed) argument 593 object_iterate_mem_careful(MemRegion mr, ObjectClosure* cl) argument 626 oops_on_card_seq_iterate_careful(MemRegion mr, FilterOutOfRegionClosure* cl, bool filter_young, jbyte* card_ptr) argument 911 initialize(MemRegion mr, bool clear_space, bool mangle_space) argument 985 G1OffsetTableContigSpace(G1BlockOffsetSharedArray* sharedOffsetArray, MemRegion mr, bool is_zeroed) argument [all...] |
/openjdk7/jdk/src/share/classes/javax/management/modelmbean/ |
H A D | ModelMBean.java | 105 * @param mr Object that is the managed resource 119 public void setManagedResource(Object mr, String mr_type) argument
|
/openjdk7/hotspot/agent/src/share/classes/sun/jvm/hotspot/gc_interface/ |
H A D | CollectedHeap.java | 82 MemRegion mr = reservedRegion(); 84 mr.start() + "," + mr.end() + ")");
|
/openjdk7/hotspot/src/share/vm/oops/ |
H A D | instanceMirrorKlass.hpp | 92 int oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) { argument 93 return oop_oop_iterate_v_m(obj, blk, mr); 98 int oop_oop_iterate##nv_suffix##_m(oop obj, OopClosureType* blk, MemRegion mr);
|
H A D | instanceRefKlass.hpp | 71 int oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) { argument 72 return oop_oop_iterate_v_m(obj, blk, mr); 77 int oop_oop_iterate##nv_suffix##_m(oop obj, OopClosureType* blk, MemRegion mr);
|
H A D | klassKlass.cpp | 130 int klassKlass::oop_oop_iterate_m(oop obj, OopClosure* blk, MemRegion mr) { argument 136 if (mr.contains(adr)) blk->do_oop(adr); 139 if (mr.contains(adr)) blk->do_oop(adr); 142 if (mr.contains(adr)) blk->do_oop(adr); 144 if (mr.contains(adr)) blk->do_oop(adr); 146 if (mr.contains(adr)) blk->do_oop(adr); 153 && (mr.contains(k->adr_subklass()) 154 || mr.contains(k->adr_next_sibling()))) { 157 obj->oop_iterate_header(blk, mr);
|
/openjdk7/jdk/test/java/io/Serializable/skipToEndOfBlockData/ |
H A D | SkipToEndOfBlockData.java | 119 MismatchedRead mr, mrcopy; 125 mr = new MismatchedRead(1, (float) 2.34); 129 oout.writeObject(mr); 140 if (! (mr.equals(mrcopy) && mre.equals(mrecopy) && ib.equals(ibcopy)))
|