Lines Matching defs:so

62 // turn it on so that the contents of the young list (scan-only /
106 // concurrently, and so we do not care if card_ptr contains references
340 // The region is a non-empty survivor so let's add it to
449 // so an address p can point to a region which will be in the
612 // just expanded the heap!) so let's just be conservative and
630 // path. The caller will attempt the expansion if this fails, so
642 // know which list they are on so that we can remove them. Note
735 // fields still point to bottom, so the intersection between
800 // Otherwise, if can expand, do so.
846 // information of the old generation so we need to recalculate the
1042 // Humongous objects can exhaust the heap quickly, so we should check if we
1366 // regions so that all the ALLOC / RETIRE events are generated
1434 // We should do this after we potentially resize the heap so
1460 // the phase so that the number of parallel threads does
1532 // TraceMemoryManagerStats is called) so that the G1 memory pools are updated
1722 // complete compaction phase than we've tried so far might be
2149 // (Must do this late, so that "max_regions" is defined.)
2221 // Do create of the monitoring and management support so that
2507 // is set) so that if a waiter requests another System.gc() it doesn't
2941 // during the current pause - so it's valid.
2949 // The cached entry was not valid so let's calculate
2974 // so let's iterate from there
3021 // The CSet is empty so there's nothing to do.
3384 const int so = SO_AllClasses | SO_Strings | SO_CodeCache;
3388 // so we don't reset the dirty cards in the perm gen.
3389 ScanningOption(so), // roots scanning options
3542 // The "G1CollectorPolicy" is keeping track of these stats, so delegate
3728 // region is freed) so they are ready when the region is
3854 // We are about to start a marking cycle, so we increment the
3877 // set, skip this step so that the region allocation code has to
3921 // region(s) so that all the ALLOC / RETIRE events are generated
3976 // we have not filtered them yet (we'll do so during the
4098 // We failed to expand the heap so let's verify that
4129 // set_saved_mark() on all the GC alloc regions, so card
4154 // We should do this after we potentially expand the heap so
4156 // event, and after we retire the GC alloc regions so that all
4190 // TraceMemoryManagerStats is called) so that the G1 memory pools are updated
4299 // want either way so no reason to check explicitly for either
4587 // Drain the overflow stack first, so other threads can steal.
4614 // We know that the object is not moving so it's safe to read its size.
4636 // worker so we cannot trust that its to-space image is
4676 // We're going to allocate linearly, so might as well prefetch ahead.
4691 // a displaced marked word, so that case is not optimized
4754 // mark it, the evacuation failure protocol will do so.
4817 // so that the heap remains parsable in case of evacuation failure.
5000 // Close the inner scope so that the ResourceMark and HandleMark
5017 // the collection set. Used as a predicate when walking code roots so
5078 ScanningOption so,
5097 collecting_perm_gen, so,
5373 // the complete_gc closure (which ordinarily would drain the queue) so
5637 // (that were not copied during the actual pause) so we can
5742 // elapsed time before closing the scope so that time
6183 // Most of the time we won't have to wait, so let's do a quick test
6560 // Make sure we append the secondary_free_list on the free_list so