Lines Matching defs:space

259   // object size (which today is a header's worth of space);
279 // space be allocated before initiating a new collection cycle.
398 // Start a cms collection if there isn't enough space to promote
756 // Support for parallelizing survivor space rescan
881 // part of the space ends in a free block we should add that to
1133 static inline size_t percent_of_space(Space* space, HeapWord* addr)
1135 size_t delta = pointer_delta(addr, space->bottom());
1136 return (size_t)(delta * 100.0 / (space->capacity() / HeapWordSize));
1152 // then compute the offset from the endpoints of the space.
1201 CMSCollector::allocation_limit_reached(Space* space, HeapWord* top,
1206 if (CMSIncrementalMode && _icms_start_limit != space->end()) {
1209 space->print_on(gclog_or_tty);
1215 percent_of_space(space, _icms_stop_limit));
1226 space->print_on(gclog_or_tty);
1231 top, space->end(),
1232 percent_of_space(space, space->end()));
1235 return space->end();
1240 space->print_on(gclog_or_tty);
1245 top, space->end(),
1246 percent_of_space(space, space->end()));
1249 return space->end();
1253 space->print_on(gclog_or_tty);
1267 // delegate to underlying space.
1307 ConcurrentMarkSweepGeneration::allocation_limit_reached(Space* space,
1311 return collector()->allocation_limit_reached(space, top, word_sz);
1369 // if we are tracking promotions, then first ensure space for
1370 // promotion (including spooling space for saving header if necessary).
1376 // Out of space for allocating spooling buffers;
1604 // . we recently needed to expand this space and have not, since that
1606 // . the underlying space believes that it may be a good idea to initiate
1608 // following: the space uses linear allocation and linear allocation is
1614 // we will be out of sufficient free space given allocation rate estimates.]
1812 // We need to lock the Free list lock for the space that we are
1965 // to reclaim as much space as possible.
2041 // in the free list space. We do so by reading and clearing
2060 "All the free space should be compacted into one chunk at top");
2064 "All the free space should be in a single chunk");
3057 // delegate to CMS space
3091 // call to enable/disable allocation profiling in a generation/space,
3093 // gen/space since the enable call, or the last iterator call (which
3360 // If there's not enough expansion space available, give up.
3368 // A competing par_promote might beat us to the expansion space,
3388 // If there's not enough expansion space available, give up.
3396 // A competing allocation might beat us to the expansion space,
3429 // update the space and generation capacity counters
3928 gclog_or_tty->print_cr("Finished cms space scanning in %dth thread: %3.3f sec",
3939 gclog_or_tty->print_cr("Finished perm space scanning in %dth thread: %3.3f sec",
4016 // for this space.
4024 // Having claimed the nth task in this space,
4039 // beyond the "top" address of the space.
4094 // more tasks left to claim in this space, the global_finger
4095 // must exceed space->top() and a fortiori space->end(). However,
4605 if (clean_survivor) { // preclean the active survivor space(s)
4691 // so that we can safely iterate over the CMS space when scanning
5148 ContiguousSpace* space,
5151 // ... of dirty cards in old space
5261 Par_MarkRefsIntoAndScanClosure* cl, ContiguousSpace* space,
5272 SequentialSubTasksDone* pst = space->par_seq_tasks();
5283 start = space->bottom();
5284 end = space->top();
5286 start = space->bottom();
5295 end = space->top();
5298 // Verify that mr is in space
5299 assert(mr.is_empty() || space->used_region().contains(mr),
5300 "Should be in space");
5304 space->par_oop_iterate(mr, cl);
5479 // array which will provide the partitioning of the survivor space
5547 // Set up the space's par_seq_tasks structure for work claiming
5556 // Eden space
5576 // To space
5587 // From space
5642 // of parallel tasks (per constituent space) that are dynamically
5682 NULL, // space is set further below
6192 gch->clear_incremental_collection_failed(); // Worth retrying as fresh space may have been freed up
6264 // We iterate over the space(s) underlying this generation,
6649 // Reinitialize virtual space for new stack
6775 assert(res, "Should have space to push on empty stack");
7313 assert(res, "Empty non-zero size stack should have space for single push");
7518 assert(res, "Empty non-zero size stack should have space for single push");
8100 // previous _end of the space), so we may have stepped past _limit:
8518 // the previous end of the space when this cycle started, and
8534 "[" PTR_FORMAT "," PTR_FORMAT ") in space "
9256 // Record the new size of the space in the cms generation
9286 // Find the chunk at the end of the space and determine
9292 "Inconsistent chunk at end of space");
9296 // Shrink the underlying space
9311 // If the space did shrink (shrinking is not guaranteed),
9342 // update the space and generation capacity counters
9357 "Inconsistency at end of space");