Lines Matching refs:HeapWord

56   HeapWord* _bottom;    // == reserved.start
57 HeapWord* _end; // End of currently allocated region.
62 BlockOffsetTable(HeapWord* bottom, HeapWord* end):
71 virtual void set_bottom(HeapWord* new_bottom) {
79 virtual HeapWord* block_start_unsafe(const void* addr) const = 0;
83 HeapWord* block_start(const void* addr) const;
123 HeapWord* _end;
147 void set_offset_array(size_t index, HeapWord* high, HeapWord* low, bool reducing = false) {
157 void set_offset_array(HeapWord* left, HeapWord* right, u_char offset, bool reducing = false) {
202 void check_offset_array(size_t index, HeapWord* high, HeapWord* low) const {
210 bool is_card_boundary(HeapWord* p) const;
238 void set_bottom(HeapWord* new_bottom);
254 HeapWord* address_for_index(size_t index) const;
260 HeapWord* inc_by_region_size(HeapWord* p) const { return p + N_words; }
263 void serialize(SerializeOopClosure* soc, HeapWord* start, HeapWord* end);
328 void set_remainder_to_point_to_start(HeapWord* start, HeapWord* end, bool reducing = false);
334 void do_block_internal(HeapWord* blk_start, HeapWord* blk_end, Action action, bool reducing = false);
359 HeapWord* new_end = _bottom + new_word_size;
375 void single_block(HeapWord* blk_start, HeapWord* blk_end);
376 void single_block(HeapWord* blk, size_t size) {
389 virtual void alloc_block(HeapWord* blk_start, HeapWord* blk_end);
390 void alloc_block(HeapWord* blk, size_t size) {
424 HeapWord* _unallocated_block;
432 HeapWord* unallocated_block() const {
438 void set_unallocated_block(HeapWord* block) {
447 void alloc_block(HeapWord* blk_start, HeapWord* blk_end);
448 void alloc_block(HeapWord* blk, size_t size) {
461 void split_block(HeapWord* blk_start, size_t full_blk_size,
469 void mark_block(HeapWord* blk_start, HeapWord* blk_end, bool reducing = false);
470 void mark_block(HeapWord* blk, size_t size, bool reducing = false) {
478 void allocated(HeapWord* blk_start, HeapWord* blk_end, bool reducing = false) {
486 void allocated(HeapWord* blk, size_t size, bool reducing = false) {
490 void freed(HeapWord* blk_start, HeapWord* blk_end);
491 void freed(HeapWord* blk, size_t size);
493 HeapWord* block_start_unsafe(const void* addr) const;
497 HeapWord* block_start_careful(const void* addr) const;
503 void verify_single_block(HeapWord* blk_start, HeapWord* blk_end)
505 void verify_single_block(HeapWord* blk, size_t size) PRODUCT_RETURN;
508 void verify_not_unallocated(HeapWord* blk_start, HeapWord* blk_end)
510 void verify_not_unallocated(HeapWord* blk, size_t size)
526 HeapWord* _next_offset_threshold;
530 void alloc_block_work(HeapWord* blk_start, HeapWord* blk_end);
542 HeapWord* initialize_threshold();
548 HeapWord* threshold() const { return _next_offset_threshold; }
555 void alloc_block(HeapWord* blk_start, HeapWord* blk_end) {
560 void alloc_block(HeapWord* blk, size_t size) {
564 HeapWord* block_start_unsafe(const void* addr) const;