/openjdk7/hotspot/src/share/vm/gc_implementation/g1/ |
H A D | g1HotCardCache.cpp | 105 int end_idx = start_idx + _hot_cache_par_chunk_size; local 108 Atomic::cmpxchg(end_idx, &_hot_cache_par_claimed_idx, start_idx)) { 109 // The current worker has successfully claimed the chunk [start_idx..end_idx) 110 end_idx = MIN2(end_idx, _n_hot); 111 for (int i = start_idx; i < end_idx; i++) {
|
H A D | concurrentMark.inline.hpp | 35 BitMap::idx_t end_idx, 38 // Set the exclusive bit range [start_idx, end_idx). 39 assert((end_idx - start_idx) > 0, "at least one card"); 40 assert(end_idx <= card_bm->size(), "sanity"); 43 end_idx = MIN2(end_idx, card_bm->size()); 49 if ((end_idx - start_idx) <= 8) { 50 for (BitMap::idx_t i = start_idx; i < end_idx; i += 1) { 60 card_bm->par_at_put_range(start_idx, end_idx, true); 62 card_bm->set_range(start_idx, end_idx); 33 set_card_bitmap_range(BitMap* card_bm, BitMap::idx_t start_idx, BitMap::idx_t end_idx, bool is_par) argument 101 BitMap::idx_t end_idx = card_bitmap_index_for(end); local [all...] |
H A D | concurrentMark.cpp | 1317 BitMap::idx_t end_idx = _cm->card_bitmap_index_for(obj_end); local 1320 // could be actually just beyond the end of the heap; end_idx 1326 end_idx += 1; 1330 _cm->set_card_bitmap_range(_card_bm, start_idx, end_idx, true /* is_par */); 1343 BitMap::idx_t end_idx = _cm->card_bitmap_index_for(top); local 1346 // could be actually just beyond the end of the heap; end_idx 1352 end_idx += 1; 1354 _cm->set_card_bitmap_range(_card_bm, start_idx, end_idx, true /* is_par */); 1469 BitMap::idx_t end_idx = _cm->card_bitmap_index_for(hr->top()); local 1471 for (BitMap::idx_t i = start_idx; i < end_idx; 1608 BitMap::idx_t end_idx = _cm->card_bitmap_index_for(top); local 2970 BitMap::idx_t end_idx = _cm->card_bitmap_index_for(end); local [all...] |
H A D | concurrentMark.hpp | 830 BitMap::idx_t end_idx,
|
/openjdk7/hotspot/src/share/vm/opto/ |
H A D | block.cpp | 167 int end_idx = _nodes.size()-1; 170 if ((end_idx > 0) && (_nodes[end_idx]->is_MachGoto())) { 172 end_idx--; 183 while ((end_idx > 0) && !_nodes[end_idx]->is_Mach()) { 184 end_idx--; 188 if (end_idx == 0) { 566 int end_idx = b->end_idx(); [all...] |
H A D | reg_split.cpp | 129 int sidx = i - b->end_idx()-1; 165 b = b->_succs[b->_nodes[b->end_idx()+1]->Opcode() == Op_IfTrue]; 274 bindex = b->end_idx(); 439 for( uint i = 1; i <= b->end_idx(); i++ ) { 630 for( insidx = 1; insidx <= b->end_idx(); insidx++ ) { 677 if( is_high_pressure( b, &lrgs(lidx), b->end_idx()) && !prompt_use(b,lidx) ) 730 for( insidx = 1; insidx <= b->end_idx(); insidx++ ) { 772 b->_ihrp_index >= b->end_idx() || 777 b->_fhrp_index >= b->end_idx() || 817 uint orig_eidx = b->end_idx(); [all...] |
H A D | lcm.cpp | 368 Node *tmp1 = _nodes[end_idx()+1]; 369 Node *tmp2 = _nodes[end_idx()+2]; 370 _nodes.map(end_idx()+1, tmp2); 371 _nodes.map(end_idx()+2, tmp1); 384 _nodes.map(end_idx(),nul_chk); 688 uint node_cnt = end_idx(); 865 if( phi_cnt != end_idx() ) { 874 // assert( phi_cnt == end_idx(), "did not schedule all" ); 1011 uint end = end_idx();
|
H A D | coalesce.cpp | 308 uint i = b->end_idx()-1; 323 uint kill_src_idx = b->end_idx(); 326 i = b->end_idx()-1; 433 _phc.clone_projs( pred, pred->end_idx(), m, copy, _phc._maxlrg ); 580 uint cnt = b->end_idx(); 665 bindex2 = b2->end_idx()-1; 907 for( uint i = 1; i<b->end_idx(); i++ ) {
|
H A D | gcm.cpp | 1019 uint end_latency = _node_latency->at_grow(LCA->_nodes[LCA->end_idx()]->_idx); 1043 LCA->_nodes[LCA->end_idx()]->_idx, 1065 uint end_idx = LCA->end_idx(); local 1066 uint end_lat = _node_latency->at_grow(LCA->_nodes[end_idx]->_idx); 1071 LCA->_pre_order, LCA->_nodes[0]->_idx, start_lat, end_idx, end_lat, LCA_freq); 1689 int eidx = end_idx(); 1773 int eidx = end_idx(); 1825 int eidx = end_idx(); 1870 int eidx = end_idx(); [all...] |
H A D | block.hpp | 261 uint end_idx() const { function in class:Block 264 // This will not give correct end_idx for block 0 when it only contains root. 274 Node *end() const { return _nodes[end_idx()]; } 278 void add_inst( Node *n ) { _nodes.insert(end_idx(),n); }
|
H A D | ifg.cpp | 343 for( uint j = b->end_idx() + 1; j > 1; j-- ) { 486 uint last_inst = b->end_idx(); 816 if( j < b->_nodes.size() && j < b->end_idx()+1 ) { 825 if( j < b->_nodes.size() && j < b->end_idx()+1 ) {
|
H A D | live.cpp | 283 for( uint j = b->end_idx() + 1; j > 1; j-- ) {
|
H A D | domgraph.cpp | 221 int eidx = b->end_idx();
|
H A D | chaitin.cpp | 1458 uint last_inst = b->end_idx(); 1600 for( i = 1; i <= b->end_idx(); i++ ) {// Search for matching Phi 1645 for( uint j = b->end_idx() + 1; j > 1; j-- ) {
|
/openjdk7/hotspot/src/share/vm/classfile/ |
H A D | symbolTable.cpp | 776 void StringTable::buckets_do(OopClosure* f, int start_idx, int end_idx) { argument 781 assert(0 <= end_idx && end_idx <= limit, 782 err_msg("end_idx (" INT32_FORMAT ") oob?", end_idx)); 783 assert(start_idx <= end_idx, 784 err_msg("Ordering: start_idx=" INT32_FORMAT", end_idx=" INT32_FORMAT, 785 start_idx, end_idx)); 787 for (int i = start_idx; i < end_idx; i += 1) { 822 int end_idx local [all...] |
H A D | symbolTable.hpp | 258 // in the range [start_idx, end_idx). 259 static void buckets_do(OopClosure* f, int start_idx, int end_idx);
|
/openjdk7/hotspot/src/share/vm/c1/ |
H A D | c1_LinearScan.cpp | 6070 int end_idx = i - 1; 6071 BlockBegin* end_block = code->at(end_idx); 6074 // short loop from header_idx to end_idx found -> reorder blocks such that 6077 end_idx - header_idx + 1, 6080 for (int j = header_idx; j < end_idx; j++) { 6083 code->at_put(end_idx, header_block); 6086 assert(code->at(end_idx)->is_set(BlockBegin::backward_branch_target_flag), "must be backward branch target"); 6087 code->at(end_idx)->clear(BlockBegin::backward_branch_target_flag);
|
/openjdk7/hotspot/src/share/vm/gc_implementation/parallelScavenge/ |
H A D | psParallelCompact.cpp | 621 const size_t end_idx = addr_to_region_idx(target_end); local 626 beg_idx, end_idx); 628 for (size_t idx = beg_idx; idx < end_idx; ++idx) {
|