Lines Matching refs:MemRegion

98   const MemRegion _whole_heap;       // the region covered by the card table
109 MemRegion* _covered;
116 MemRegion* _committed;
120 // uncommit the MemRegion for that page.
121 MemRegion _guard_region;
140 void resize_covered_region(MemRegion new_region);
151 MemRegion committed_unique_to_self(int self, MemRegion mr) const;
176 void non_clean_card_iterate_serial(MemRegion mr, MemRegionClosure* cl);
184 void non_clean_card_iterate_possibly_parallel(Space* sp, MemRegion mr,
190 void non_clean_card_iterate_parallel_work(Space* sp, MemRegion mr,
197 void dirty_MemRegion(MemRegion mr);
201 void clear_MemRegion(MemRegion mr);
228 size_t chunks_to_cover(MemRegion mr) {
243 MemRegion used,
256 MemRegion chunk_mr,
257 MemRegion used,
282 CardTableModRefBS(MemRegion whole_heap, int max_covered_regions);
306 inline void inline_write_region(MemRegion mr) {
310 void write_region_work(MemRegion mr) {
315 inline void inline_write_ref_array(MemRegion mr) {
319 void write_ref_array_work(MemRegion mr) {
404 virtual void invalidate(MemRegion mr, bool whole_heap = false);
405 void clear(MemRegion mr);
406 void dirty(MemRegion mr);
412 // MemRegion). Thus, several modified cards may be lumped into one
422 // for cards within the MemRegion "mr" (which is required to be
424 void mod_card_iterate(MemRegion mr, MemRegionClosure* cl) {
431 // MemRegion "mr".
432 void dirty_card_iterate(MemRegion mr, MemRegionClosure* cl);
434 // Return the MemRegion corresponding to the first maximal run
435 // of dirty cards lying completely within MemRegion mr.
438 MemRegion dirty_card_range_after_reset(MemRegion mr, bool reset,
483 void verify_region(MemRegion mr, jbyte val, bool val_equals) PRODUCT_RETURN;
484 void verify_not_dirty_region(MemRegion mr) PRODUCT_RETURN;
485 void verify_dirty_region(MemRegion mr) PRODUCT_RETURN;
502 CardTableModRefBSForCTRS(MemRegion whole_heap,