gcm.cpp revision 2958
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * Copyright (c) 1997, 2011, Oracle and/or its affiliates. All rights reserved.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * This code is free software; you can redistribute it and/or modify it
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * under the terms of the GNU General Public License version 2 only, as
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * published by the Free Software Foundation.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * This code is distributed in the hope that it will be useful, but WITHOUT
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * version 2 for more details (a copy is included in the LICENSE file that
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * accompanied this code).
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * You should have received a copy of the GNU General Public License version
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * 2 along with this work; if not, write to the Free Software Foundation,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift * or visit www.oracle.com if you need additional information or have any
2e334e4f58d7760ed0788f5db5524d87934d6869matthew_swift// Portions of code courtesy of Clifford Click
2e334e4f58d7760ed0788f5db5524d87934d6869matthew_swift// Optimization - Graph Style
2e334e4f58d7760ed0788f5db5524d87934d6869matthew_swift// To avoid float value underflow
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift//----------------------------schedule_node_into_block-------------------------
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift// Insert node n into block b. Look for projections of n and make sure they
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift// are in b also.
2e334e4f58d7760ed0788f5db5524d87934d6869matthew_swiftvoid PhaseCFG::schedule_node_into_block( Node *n, Block *b ) {
2e334e4f58d7760ed0788f5db5524d87934d6869matthew_swift // Set basic block of n, Add n to b,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // After Matching, nearly any old Node may have projections trailing it.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // These are usually machine-dependent flags. In any case, they might
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // float to another block below this one. Move them up.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (DUIterator_Fast imax, i = n->fast_outs(imax); i < imax; i++) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift buse->find_remove(use); // Remove from wrong block
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift _bbs.map(use->_idx, b); // Re-insert in this block
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//----------------------------replace_block_proj_ctrl-------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Nodes that have is_block_proj() nodes as their control need to use
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// the appropriate Region for their actual block as their control since
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// the projection will be in a predecessor block.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftvoid PhaseCFG::replace_block_proj_ctrl( Node *n ) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(in0 != NULL, "Only control-dependent");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (p != NULL && p != n) { // Control from a block projection?
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift assert(!n->pinned() || n->is_MachConstantBase(), "only pinned MachConstantBase node is expected here");
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift // Find trailing Region
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift Block *pb = _bbs[in0->_idx]; // Block-projection already has basic block
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift if (pb->_num_succs != 1) { // More then 1 successor?
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift // Search for successor
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift // Find which output path belongs to projection
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift // Change control to match head of successor basic block
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift//------------------------------schedule_pinned_nodes--------------------------
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift// Set the basic block for Nodes pinned into blocks
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swiftvoid PhaseCFG::schedule_pinned_nodes( VectorSet &visited ) {
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift // Allocate node stack of size C->unique()+8 to avoid frequent realloc
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift GrowableArray <Node *> spstack(C->unique()+8);
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift if( !visited.test_set(n->_idx) ) { // Test node and flag it as visited
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift if( n->pinned() && !_bbs.lookup(n->_idx) ) { // Pinned? Nail it down!
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift assert( n->in(0), "pinned Node must have Control" );
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift // Before setting block replace block_proj control edge
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift Block *b = _bbs[input->_idx]; // Basic block of controlling input
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift for( int i = n->req() - 1; i >= 0; --i ) { // For all inputs
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift// Assert that new input b2 is dominated by all previous inputs.
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift// Check this by by seeing that it is dominated by b1, the deepest
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift// input observed until b2.
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swiftstatic void assert_dom(Block* b1, Block* b2, Node* n, Block_Array &bbs) {
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift assert(b1->_dom_depth < b2->_dom_depth, "sanity");
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift // Detected an unschedulable graph. Print some nice stuff and die.
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift for (uint j=0; j<n->len(); j++) { // For all inputs
0c45338572f2f6593fea8fc9e70bcb9ef3a6d2f8matthew_swift if (inn == NULL) continue; // Ignore NULL, missing inputs
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print("B%d idom=B%d depth=%2d ",inb->_pre_order,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift inb->_idom ? inb->_idom->_pre_order : 0, inb->_dom_depth);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftstatic Block* find_deepest_input(Node* n, Block_Array &bbs) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Find the last input dominated by all other inputs.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (uint k = 0; k < n->len(); k++) { // For all inputs
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (inn == NULL) continue; // Ignore NULL, missing inputs
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(inb != NULL, "must already have scheduled this input");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (deepb_dom_depth < (int) inb->_dom_depth) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // The new inb must be dominated by the previous deepb.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // The various inputs must be linearly ordered in the dom
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // tree, or else there will not be a unique deepest block.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(deepb != NULL, "must be at least one input to n");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------schedule_early---------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Find the earliest Block any instruction can be placed in. Some instructions
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// are pinned into Blocks. Unpinned instructions can appear in last block in
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// which all their inputs occur.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftbool PhaseCFG::schedule_early(VectorSet &visited, Node_List &roots) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Allocate stack with enough space to avoid frequent realloc
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift Node_Stack nstack(roots.Size() + 8); // (unique >> 1) + 24 from Java2D stats
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // roots.push(_root); _root will be processed among C->top() inputs
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // Use local variables nstack_top_n & nstack_top_i to cache values
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // on stack's top.
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift//while_nstack_nonempty:
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift while (true) {
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // Get parent node and next input's index from stack's top.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (i == 0) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Fixup some control. Constants without control get attached
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // to root and nodes that use is_block_proj() nodes should be attached
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // to the region that starts their block.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift } else { // n->in(0) == NULL
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (n->req() == 1) { // This guy is a constant with NO inputs?
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // First, visit all inputs and force them to get a block. If an
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // input is already in a block we quit following inputs (to avoid
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // cycles). Instead we put that Node on a worklist to be handled
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // later (since IT'S inputs may not have a block yet).
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift bool done = true; // Assume all n's inputs will be processed
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (in == NULL) continue; // Ignore NULL, missing inputs
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (!_bbs.lookup(in->_idx)) { // Missing block selection?
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // assert( !visited.test(in->_idx), "did not schedule early" );
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift return false;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift nstack.push(n, i); // Save parent node and next input's index.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift nstack_top_n = in; // Process current input now.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift done = false; // Not all n's inputs processed.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift break; // continue while_nstack_nonempty;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift } else if (!is_visited) { // Input not yet visited?
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift roots.push(in); // Visit this guy later, using worklist
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // All of n's inputs have been processed, complete post-processing.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Some instructions are pinned into a block. These include Region,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Phi, Start, Return, and other control-dependent instructions and
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // any projections which depend on them.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Set earliest legal block.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift _bbs.map(n->_idx, find_deepest_input(n, _bbs));
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(_bbs[n->_idx] == _bbs[n->in(0)->_idx], "Pinned Node should be at the same block as its control edge");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Finished all nodes on stack.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Process next node on the worklist 'roots'.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Get saved parent node and next input's index.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift } // if (done)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift } // while (true)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift } // while (roots.size() != 0)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------dom_lca----------------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Find least common ancestor in dominator tree
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// LCA is a current notion of LCA, to be raised above 'this'.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// As a convenient boundary condition, return 'this' if LCA is NULL.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Find the LCA of those two nodes.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift anc = anc->_idom; // Walk up till anc is as high as LCA
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift LCA = LCA->_idom; // Walk up till LCA is as high as anc
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift while (LCA != anc) { // Walk both up till they are the same
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//--------------------------raise_LCA_above_use--------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// We are placing a definition, and have been given a def->use edge.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// The definition must dominate the use, so move the LCA upward in the
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// dominator tree to dominate the use. If the use is a phi, adjust
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// the LCA only with the phi input paths which actually use this def.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftstatic Block* raise_LCA_above_use(Block* LCA, Node* use, Node* def, Block_Array &bbs) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (buse == NULL) return LCA; // Unused killing Projs have no use block
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (!use->is_Phi()) return buse->dom_lca(LCA);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift uint pmax = use->req(); // Number of Phi inputs
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Why does not this loop just break after finding the matching input to
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // the Phi? Well...it's like this. I do not have true def-use/use-def
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // chains. Means I cannot distinguish, from the def-use direction, which
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // of many use-defs lead from the same use to the same def. That is, this
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Phi might have several uses of the same def. Each use appears in a
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // different predecessor block. But when I enter here, I cannot distinguish
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // which use-def edge I should find the predecessor block for. So I find
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // them all. Means I do a little extra work if a Phi uses the same value
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // more than once.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (uint j=1; j<pmax; j++) { // For all inputs
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (use->in(j) == def) { // Found matching input?
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//----------------------------raise_LCA_above_marks----------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Return a new LCA that dominates LCA and any of its marked predecessors.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Search all my parents up to 'early' (exclusive), looking for predecessors
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// which are marked with the given index. Return the LCA (in the dom tree)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// of all marked blocks. If there are none marked, return the original
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftstatic Block* raise_LCA_above_marks(Block* LCA, node_idx_t mark,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (mid == early) continue; // stop searching here
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Test and set the visited bit.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (mid->raise_LCA_visited() == mark) continue; // already visited
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Don't process the current LCA, otherwise the search may terminate early
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (mid != LCA && mid->raise_LCA_mark() == mark) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Raise the LCA.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (LCA == early) break; // stop searching everywhere
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(early->dominates(LCA), "early is high enough");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Resume searching at that point, skipping intermediate levels.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift continue; // Don't mark as visited to avoid early termination.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Keep searching through this block's predecessors.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (uint j = 1, jmax = mid->num_preds(); j < jmax; j++) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift Block* mid_parent = bbs[ mid->pred(j)->_idx ];
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//--------------------------memory_early_block--------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// This is a variation of find_deepest_input, the heart of schedule_early.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Find the "early" block for a load, if we considered only memory and
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// address inputs, that is, if other data inputs were ignored.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Because a subset of edges are considered, the resulting block will
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// be earlier (at a shallower dom_depth) than the true schedule_early
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// point of the node. We compute this earlier block as a more permissive
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// site for anti-dependency insertion, but only if subsume_loads is enabled.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftstatic Block* memory_early_block(Node* load, Block* early, Block_Array &bbs) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(base != NodeSentinel && index != NodeSentinel,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift "unexpected base/index inputs");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (base != NULL) mem_inputs[mem_inputs_length++] = base;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (index != NULL) mem_inputs[mem_inputs_length++] = index;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (store != NULL) mem_inputs[mem_inputs_length++] = store;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // In the comparision below, add one to account for the control input,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // which may be null, but always takes up a spot in the in array.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (mem_inputs_length + 1 < (int) load->req()) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // This "load" has more inputs than just the memory, base and index inputs.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // For purposes of checking anti-dependences, we need to start
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // from the early block of only the address portion of the instruction,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // and ignore other blocks that may have factored into the wider
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // schedule_early calculation.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (load->in(0) != NULL) mem_inputs[mem_inputs_length++] = load->in(0);
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift for (int i = 0; i < mem_inputs_length; i++) {
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift if (deepb_dom_depth < (int) inb->_dom_depth) {
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift // The new inb must be dominated by the previous deepb.
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift // The various inputs must be linearly ordered in the dom
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift // tree, or else there will not be a unique deepest block.
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift DEBUG_ONLY(assert_dom(deepb, inb, load, bbs));
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//--------------------------insert_anti_dependences---------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// A load may need to witness memory that nearby stores can overwrite.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// For each nearby store, either insert an "anti-dependence" edge
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// from the load to the store, or else move LCA upward to force the
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// load to (eventually) be scheduled in a block above the store.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Do not add edges to stores on distinct control-flow paths;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// only add edges to stores which might interfere.
71b64b6e7ff616e542de7e3fa1680fe9ca640712matthew_swift// Return the (updated) LCA. There will not be any possibly interfering
71b64b6e7ff616e542de7e3fa1680fe9ca640712matthew_swift// store between the load's "early block" and the updated LCA.
71b64b6e7ff616e542de7e3fa1680fe9ca640712matthew_swift// Any stores in the updated LCA will have new precedence edges
71b64b6e7ff616e542de7e3fa1680fe9ca640712matthew_swift// back to the load. The caller is expected to schedule the load
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// in the LCA, in which case the precedence edges will make LCM
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// preserve anti-dependences. The caller may also hoist the load
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// above the LCA, if it is not the early block.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftBlock* PhaseCFG::insert_anti_dependences(Block* LCA, Node* load, bool verify) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(load->needs_anti_dependence_check(), "must be a load of some sort");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Compute the alias index. Loads and stores with different alias indices
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // do not need anti-dependence edges.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift uint load_alias_idx = C->get_alias_index(load->adr_type());
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (load_alias_idx == Compile::AliasIdxBot && C->AliasLevel() > 0 &&
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift PrintMiscellaneous && (WizardMode || Verbose))) {
ce307551b126a6b551358fb063614b80b9355907matthew_swift // Load nodes should not consume all of memory.
ce307551b126a6b551358fb063614b80b9355907matthew_swift // Reporting a bottom type indicates a bug in adlc.
ce307551b126a6b551358fb063614b80b9355907matthew_swift // If some particular type of node validly consumes all of memory,
ce307551b126a6b551358fb063614b80b9355907matthew_swift // sharpen the preceding "if" to exclude it, so we can catch bugs here.
ce307551b126a6b551358fb063614b80b9355907matthew_swift tty->print_cr("*** Possible Anti-Dependence Bug: Load consumes all of memory.");
ce307551b126a6b551358fb063614b80b9355907matthew_swift if (VerifyAliases) assert(load_alias_idx != Compile::AliasIdxBot, "");
ce307551b126a6b551358fb063614b80b9355907matthew_swift assert(load_alias_idx || (load->is_Mach() && load->as_Mach()->ideal_Opcode() == Op_StrComp),
ce307551b126a6b551358fb063614b80b9355907matthew_swift "String compare is only known 'load' that does not conflict with any stores");
ce307551b126a6b551358fb063614b80b9355907matthew_swift assert(load_alias_idx || (load->is_Mach() && load->as_Mach()->ideal_Opcode() == Op_StrEquals),
ce307551b126a6b551358fb063614b80b9355907matthew_swift "String equals is a 'load' that does not conflict with any stores");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(load_alias_idx || (load->is_Mach() && load->as_Mach()->ideal_Opcode() == Op_StrIndexOf),
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift "String indexOf is a 'load' that does not conflict with any stores");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(load_alias_idx || (load->is_Mach() && load->as_Mach()->ideal_Opcode() == Op_AryEq),
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift "Arrays equals is a 'load' that do not conflict with any stores");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (!C->alias_type(load_alias_idx)->is_rewritable()) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // It is impossible to spoil this load by putting stores before it,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // because we know that the stores will never update the value
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // which 'load' must witness.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Note the earliest legal placement of 'load', as determined by
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // by the unique point in the dom tree where all memory effects
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // and other inputs are first available. (Computed by schedule_early.)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // For normal loads, 'early' is the shallowest place (dom graph wise)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // to look for anti-deps between this load and any store.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // If we are subsuming loads, compute an "early" block that only considers
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // memory or address inputs. This block may be different than the
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // schedule_early block in that it could be at an even shallower depth in the
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // dominator tree, and allow for a broader discovery of anti-dependences.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift early = memory_early_block(load, early, _bbs);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift ResourceArea *area = Thread::current()->resource_area();
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift Node_List worklist_mem(area); // prior memory state to store
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift Node_List worklist_store(area); // possible-def to explore
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift Node_List worklist_visited(area); // visited mergemem nodes
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift Node_List non_early_stores(area); // all relevant stores outside of early
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // %%% This extra checking fails because MergeMem nodes are not GVNed.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Provide "phi_inputs" to check if every input to a PhiNode is from the
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // original memory state. This indicates a PhiNode for which should not
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // prevent the load from sinking. For such a block, set_raise_LCA_mark
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // may be overly conservative.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Mechanism: count inputs seen for each Phi encountered in worklist_store.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift DEBUG_ONLY(GrowableArray<uint> phi_inputs(area, C->unique(),0,0));
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // 'load' uses some memory state; look for users of the same state.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Recurse through MergeMem nodes to the stores that use them.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Each of these stores is a possible definition of memory
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // that 'load' needs to use. We need to force 'load'
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // to occur before each such store. When the store is in
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // the same block as 'load', we insert an anti-dependence
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // edge load->store.
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // The relevant stores "nearby" the load consist of a tree rooted
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // at initial_mem, with internal nodes of type MergeMem.
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // Therefore, the branches visited by the worklist are of this form:
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // initial_mem -> (MergeMem ->)* store
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // The anti-dependence constraints apply only to the fringe of this tree.
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift Node* initial_mem = load->in(MemNode::Memory);
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // Examine a nearby store to see if it might interfere with our load.
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // MergeMems do not directly have anti-deps.
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // Treat them as internal nodes in a forward tree of memory states,
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // the leaves of which are each a 'possible-def'.
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift if (store == initial_mem // root (exclusive) of tree we are searching
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift || op == Op_MergeMem // internal node of tree we are searching
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift mem = store; // It's not a possibly interfering store.
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift initial_mem = NULL; // only process initial memory once
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift for (DUIterator_Fast imax, i = mem->fast_outs(imax); i < imax; i++) {
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // Be sure we don't get into combinatorial problems.
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // (Allow phis to be repeated; they can merge two relevant states.)
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift for (; j > 0; j--) {
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift if (j > 0) continue; // already on work list; do not repeat
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift if (op == Op_MachProj || op == Op_Catch) continue;
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift if (store->needs_anti_dependence_check()) continue; // not really a store
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // Compute the alias index. Loads and stores with different alias
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // indices do not need anti-dependence edges. Wide MemBar's are
91940c5122d22be4fd20bc41db2ffc4a9ba4ce59matthew_swift // anti-dependent on everything (except immutable memories).
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (!C->can_alias(adr_type, load_alias_idx)) continue;
188a85993cf1cf9925338176e5f27b95a5891c50matthew_swift // Most slow-path runtime calls do NOT modify Java memory, but
188a85993cf1cf9925338176e5f27b95a5891c50matthew_swift // they can block and so write Raw memory.
188a85993cf1cf9925338176e5f27b95a5891c50matthew_swift // Check for call into the runtime using the Java calling
188a85993cf1cf9925338176e5f27b95a5891c50matthew_swift // convention (and from there into a wrapper); it has no
188a85993cf1cf9925338176e5f27b95a5891c50matthew_swift // _method. Can't do this optimization for Native calls because
188a85993cf1cf9925338176e5f27b95a5891c50matthew_swift // they CAN write to Java memory.
188a85993cf1cf9925338176e5f27b95a5891c50matthew_swift if (mstore->ideal_Opcode() == Op_CallStaticJava) {
188a85993cf1cf9925338176e5f27b95a5891c50matthew_swift MachSafePointNode* ms = (MachSafePointNode*) mstore;
188a85993cf1cf9925338176e5f27b95a5891c50matthew_swift MachCallJavaNode* mcj = (MachCallJavaNode*) ms;
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // These runtime calls do not write to Java visible memory
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // (other than Raw) and so do not require anti-dependence edges.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // Same for SafePoints: they read/write Raw but only read otherwise.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // This is basically a workaround for SafePoints only defining control
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // instead of control + memory.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // Some raw memory, such as the load of "top" at an allocation,
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // can be control dependent on the previous safepoint. See
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // comments in GraphKit::allocate_heap() about control input.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // Inserting an anti-dep between such a safepoint and a use
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // creates a cycle, and will cause a subsequent failure in
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // local scheduling. (BugId 4919904)
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // (%%% How can a control input be a safepoint and not a projection??)
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift if (mstore->ideal_Opcode() == Op_SafePoint && load->in(0) == mstore)
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // Identify a block that the current load must be above,
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // or else observe that 'store' is all the way up in the
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // earliest legal block for 'load'. In the latter case,
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // immediately insert an anti-dependence edge.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift assert(store_block != NULL, "unused killing projections skipped above");
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // 'load' uses memory which is one (or more) of the Phi's inputs.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // It must be scheduled not before the Phi, but rather before
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // each of the relevant Phi inputs.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // Instead of finding the LCA of all inputs to a Phi that match 'mem',
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // we mark each corresponding predecessor block and do a combined
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // hoisting operation later (raise_LCA_above_marks).
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // Do not assert(store_block != early, "Phi merging memory after access")
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // PhiNode may be at start of block 'early' with backedge to 'early'
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift for (uint j = PhiNode::Input, jmax = store->req(); j < jmax; j++) {
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift if (store->in(j) == mem) { // Found matching input?
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift Block* pred_block = _bbs[store_block->pred(j)->_idx];
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // If any predecessor of the Phi matches the load's "early block",
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // we do not need a precedence edge between the Phi and 'load'
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // since the load will be forced into a block preceding the Phi.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift early->dominates(pred_block), "early is high enough");
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // anti-dependent upon PHI pinned below 'early', no edge needed
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift LCA = early; // but can not schedule below 'early'
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // This assert asks about correct handling of PhiNodes, which may not
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // have all input edges directly from 'mem'. See BugId 4621264
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift int num_mem_inputs = phi_inputs.at_grow(store->_idx,0) + 1;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Increment by exactly one even if there are multiple copies of 'mem'
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // coming into the phi, because we will run this block several times
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // if there are several copies of 'mem'. (That's how DU iterators work.)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift phi_inputs.at_put(store->_idx, num_mem_inputs);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(PhiNode::Input + num_mem_inputs < store->req(),
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift "Expect at least one phi input will not be from original memory state");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift#endif //ASSERT
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift#endif //TRACK_PHI_INPUTS
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // 'store' is between the current LCA and earliest possible block.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Label its block, and decide later on how to raise the LCA
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // to include the effect on LCA of this store.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // If this store's block gets chosen as the raised LCA, we
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // will find him on the non_early_stores list and stick him
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // with a precedence edge.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // (But, don't bother if LCA is already raised all the way.)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Found a possibly-interfering store in the load's 'early' block.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // This means 'load' cannot sink at all in the dominator tree.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Add an anti-dep edge, and squeeze 'load' into the highest block.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(store != load->in(0), "dependence cycle found");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(store->find_edge(load) != -1, "missing precedence edge");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // This turns off the process of gathering non_early_stores.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // (Worklist is now empty; all nearby stores have been visited.)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Finished if 'load' must be scheduled in its 'early' block.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // If we found any stores there, they have already been given
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // precedence edges.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // We get here only if there are no possibly-interfering stores
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // in the load's 'early' block. Move LCA up above all predecessors
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // which contain stores we have noted.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // The raised LCA block can be a home to such interfering stores,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // but its predecessors must not contain any such stores.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // The raised LCA will be a lower bound for placing the load,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // preventing the load from sinking past any block containing
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // a store that may invalidate the memory state required by 'load'.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift LCA = raise_LCA_above_marks(LCA, load->_idx, early, _bbs);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Insert anti-dependence edges from 'load' to each store
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // in the non-early LCA block.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Mine the non_early_stores list for such stores.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // add anti_dependence from store to load in its own block
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift assert(store != load->in(0), "dependence cycle found");
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift assert(store->find_edge(load) != -1, "missing precedence edge");
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift assert(store_block->raise_LCA_mark() == load_index, "block was marked");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Any other stores we found must be either inside the new LCA
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // or else outside the original LCA. In the latter case, they
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // did not interfere with any use of 'load'.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift || !LCA_orig->dominates(store_block), "no stray stores");
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // Return the highest block containing stores; any stores
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // within that block have been given anti-dependence edges.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift// This class is used to iterate backwards over the nodes in the graph.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Constructor for the iterator
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift Node_Backward_Iterator(Node *root, VectorSet &visited, Node_List &stack, Block_Array &bbs);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Postincrement operator to iterate over the nodes
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift// Constructor for the Node_Backward_Iterator
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swiftNode_Backward_Iterator::Node_Backward_Iterator( Node *root, VectorSet &visited, Node_List &stack, Block_Array &bbs )
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift : _visited(visited), _stack(stack), _bbs(bbs) {
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // The stack should contain exactly the root
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // Clear the visited bits
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift// Iterator for the Node_Backward_Iterator
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // If the _stack is empty, then just return NULL: finished.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // '_stack' is emulating a real _stack. The 'visit-all-users' loop has been
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // made stateless, so I do not need to record the index 'i' on my _stack.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // Instead I visit all users each time, scanning for unvisited users.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // I visit unvisited not-anti-dependence users first, then anti-dependent
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // children next.
7edb1a6332a5678c67bfa236b629d47cab1525c1matthew_swift // I cycle here when I am entering a deeper level of recursion.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // The key variable 'self' was set prior to jumping here.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Now schedule all uses as late as possible.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift uint src = self->is_Proj() ? self->in(0)->_idx : self->_idx;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Schedule all nodes in a post-order visit
c9020c301eada7085bb81da1821cac5533a1f20fmatthew_swift Node *unvisited = NULL; // Unvisited anti-dependent Node, if any
c9020c301eada7085bb81da1821cac5533a1f20fmatthew_swift // Scan for unvisited nodes
c9020c301eada7085bb81da1821cac5533a1f20fmatthew_swift for (DUIterator_Fast imax, i = self->fast_outs(imax); i < imax; i++) {
c9020c301eada7085bb81da1821cac5533a1f20fmatthew_swift // For all uses, schedule late
c9020c301eada7085bb81da1821cac5533a1f20fmatthew_swift // Skip already visited children
c9020c301eada7085bb81da1821cac5533a1f20fmatthew_swift // do not traverse backward control edges
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Phi nodes always precede uses in a basic block
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Check for possible-anti-dependent
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift break; // Not visited, not anti-dep; schedule it NOW
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Did I find an unvisited not-anti-dependent Node?
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift break; // All done with children; post-visit 'self'
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Visit the unvisited Node. Contains the obvious push to
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // indicate I'm entering a deeper level of recursion. I push the
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // old state onto the _stack and set a new state and loop (recurse).
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift } // End recursion loop
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------ComputeLatenciesBackwards----------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Compute the latency of all the instructions.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftvoid PhaseCFG::ComputeLatenciesBackwards(VectorSet &visited, Node_List &stack) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print("\n#---- ComputeLatenciesBackwards ----\n");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift Node_Backward_Iterator iter((Node *)_root, visited, stack, _bbs);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Walk over all the nodes from last to first
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Set the latency for the definitions of this instruction
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift} // end ComputeLatenciesBackwards
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------partial_latency_of_defs------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Compute the latency impact of this node on all defs. This computes
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// a number that increases as we approach the beginning of the routine.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftvoid PhaseCFG::partial_latency_of_defs(Node *n) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Set the latency for this instruction
9355c1819517672f3cfe09356402fe0257c96facmatthew_swift tty->print("# latency_to_inputs: node_latency[%d] = %d for node",
cd827d0bd7c904782575c994f30982ffe6e696d5matthew_swift uint use_latency = _node_latency->at_grow(n->_idx);
cd827d0bd7c904782575c994f30982ffe6e696d5matthew_swift uint use_pre_order = _bbs[n->_idx]->_pre_order;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Walk backwards thru projections
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // If the defining block is not known, assume it is ok
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift uint def_pre_order = def_block ? def_block->_pre_order : 0;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift (use_pre_order == def_pre_order && n->is_Phi()) )
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift uint current_latency = delta_latency + use_latency;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (_node_latency->at_grow(def->_idx) < current_latency) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift _node_latency->at_put_grow(def->_idx, current_latency);
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift tty->print_cr("# %d + edge_latency(%d) == %d -> %d, node_latency[%d] = %d",
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift use_latency, j, delta_latency, current_latency, def->_idx,
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift//------------------------------latency_from_use-------------------------------
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift// Compute the latency of a specific use
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swiftint PhaseCFG::latency_from_use(Node *n, const Node *def, Node *use) {
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift // If self-reference, return no latency
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift uint def_pre_order = _bbs[def->_idx]->_pre_order;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // If the use is not a projection, then it is simple...
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift uint use_pre_order = _bbs[use->_idx]->_pre_order;
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift if (use_pre_order == def_pre_order && use->is_Phi())
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift // Change this if we want local latencies
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print_cr("# %d + edge_latency(%d) == %d -> %d, latency = %d",
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift // This is a projection, just grab the latency of the use(s)
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift for (DUIterator_Fast jmax, j = use->fast_outs(jmax); j < jmax; j++) {
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift uint l = latency_from_use(use, def, use->fast_out(j));
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift//------------------------------latency_from_uses------------------------------
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift// Compute the latency of this instruction relative to all of it's uses.
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift// This computes a number that increases as we approach the beginning of the
3cbc49fe5e6d8a00a69fe1859874ed5dcff85aebmatthew_swift // Set the latency for this instruction
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print("# latency_from_outputs: node_latency[%d] = %d for node",
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (DUIterator_Fast imax, i = n->fast_outs(imax); i < imax; i++) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift uint l = latency_from_use(n, def, n->fast_out(i));
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------hoist_to_cheaper_block-------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Pick a block for node self, between early and LCA, that is a cheaper
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// alternative to LCA.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftBlock* PhaseCFG::hoist_to_cheaper_block(Block* LCA, Block* early, Node* self) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift uint target = _node_latency->at_grow(self->_idx);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift uint start_latency = _node_latency->at_grow(LCA->_nodes[0]->_idx);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift uint end_latency = _node_latency->at_grow(LCA->_nodes[LCA->end_idx()]->_idx);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Turn off latency scheduling if scheduling is just plain off
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Do not hoist (to cover latency) instructions which target a
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // single register. Hoisting stretches the live range of the
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // single register and may force spilling.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift MachNode* mach = self->is_Mach() ? self->as_Mach() : NULL;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (mach && mach->out_RegMask().is_bound1() && mach->out_RegMask().is_NotEmpty())
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print("# Find cheaper block for latency %d: ",
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print_cr("# B%d: start latency for [%4d]=%d, end latency for [%4d]=%d, freq=%g",
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Walk up the dominator tree from LCA (Lowest common ancestor) to
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // the earliest legal location. Capture the least execution frequency.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift LCA = LCA->_idom; // Follow up the dominator tree
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Bailout without retry
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift C->record_method_not_compilable("late schedule failed: LCA == NULL");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Don't hoist machine instructions to the root basic block
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift uint start_lat = _node_latency->at_grow(LCA->_nodes[0]->_idx);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift uint end_lat = _node_latency->at_grow(LCA->_nodes[end_idx]->_idx);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print_cr("# B%d: start latency for [%4d]=%d, end latency for [%4d]=%d, freq=%g",
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift LCA->_pre_order, LCA->_nodes[0]->_idx, start_lat, end_idx, end_lat, LCA_freq);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (LCA_freq < least_freq || // Better Frequency
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift ( !in_latency && // No block containing latency
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift LCA_freq < least_freq * delta && // No worse frequency
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift !self->is_iteratively_computed() ) // But don't hoist IV increments
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // because they may end up above other uses of their phi forcing
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // their result register to be different from their input.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print_cr("# Choose block B%d with start latency=%d and freq=%g",
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift least->_pre_order, start_latency, least_freq);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // See if the latency needs to be updated
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print_cr("# Change latency for [%4d] from %d to %d", self->_idx, target, end_latency);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift _node_latency->at_put_grow(self->_idx, end_latency);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------schedule_late-----------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Now schedule all codes as LATE as possible. This is the LCA in the
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// dominator tree of all USES of a value. Pick the block with the least
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// loop nesting depth that is lowest in the dominator tree.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftextern const char must_clone[];
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftvoid PhaseCFG::schedule_late(VectorSet &visited, Node_List &stack) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift Node_Backward_Iterator iter((Node *)_root, visited, stack, _bbs);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Walk over all the nodes from last to first
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift Block* early = _bbs[self->_idx]; // Earliest legal placement
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Top node goes in bb #2 with other constants.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // It must be special-cased, because it has no out edges.
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift // No uses, just terminate
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift continue; // Must be a dead machine projection
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift // If node is pinned in the block, then no scheduling can be done.
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift MachNode* mach = self->is_Mach() ? self->as_Mach() : NULL;
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift // Don't move exception creation
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Don't move CheckCastPP nodes away from their input, if the input
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // is a rawptr (5071820).
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (def != NULL && def->bottom_type()->base() == Type::RawPtr) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Gather LCA of all uses
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift for (DUIterator_Fast imax, i = self->fast_outs(imax); i < imax; i++) {
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift // For all uses, find LCA
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift LCA = raise_LCA_above_use(LCA, use, self, _bbs);
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift } // (Hide defs of imax, i from rest of block.)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Place temps in the block of their use. This isn't a
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // requirement for correctness but it reduces useless
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // interference between temps and other nodes.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Check if 'self' could be anti-dependent on memory
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Hoist LCA above possible-defs and insert anti-dependences to
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // defs in new LCA block.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Somehow the LCA has moved above the earliest legal point.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // (One way this can happen is via memory_early_block.)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (C->subsume_loads() == true && !C->failing()) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Retry with subsume_loads == false
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // If this is the first failure, the sentinel string will "stick"
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // to the Compile object, and the C2Compiler will see it and retry.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift C->record_failure(C2Compiler::retry_no_subsuming_loads());
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // Bailout without retry when (early->_dom_depth > LCA->_dom_depth)
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift C->record_method_not_compilable("late schedule failed: incorrect graph");
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // If there is no opportunity to hoist, then we're done.
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // Must clone guys stay next to use; no hoisting allowed.
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // Also cannot hoist guys that alter memory or are otherwise not
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // allocatable (hoisting can make a value live longer, leading to
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // anti and output dependency problems which are normally resolved
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // by the register allocator giving everyone a different register).
9860f501d4514087515a9b3af4aded9876092276matthew_swift if (mach != NULL && must_clone[mach->ideal_Opcode()])
f65733ef5ab8cedafc18f3f1fb55892fb4a6ad24matthew_swift // Now find the block with the least execution frequency.
f65733ef5ab8cedafc18f3f1fb55892fb4a6ad24matthew_swift // Start at the latest schedule and work up to the earliest schedule
f65733ef5ab8cedafc18f3f1fb55892fb4a6ad24matthew_swift // in the dominator tree. Thus the Node will dominate all its uses.
f65733ef5ab8cedafc18f3f1fb55892fb4a6ad24matthew_swift late = hoist_to_cheaper_block(LCA, early, self);
9860f501d4514087515a9b3af4aded9876092276matthew_swift // Just use the LCA of the uses.
f65733ef5ab8cedafc18f3f1fb55892fb4a6ad24matthew_swift // Put the node into target block
9860f501d4514087515a9b3af4aded9876092276matthew_swift // since precedence edges are only inserted when we're sure they
9860f501d4514087515a9b3af4aded9876092276matthew_swift // are needed make sure that after placement in a block we don't
f65733ef5ab8cedafc18f3f1fb55892fb4a6ad24matthew_swift // need any new precedence edges.
9860f501d4514087515a9b3af4aded9876092276matthew_swift } // Loop until all nodes have been visited
f65733ef5ab8cedafc18f3f1fb55892fb4a6ad24matthew_swift} // end ScheduleLate
f65733ef5ab8cedafc18f3f1fb55892fb4a6ad24matthew_swift//------------------------------GlobalCodeMotion-------------------------------
f65733ef5ab8cedafc18f3f1fb55892fb4a6ad24matthew_swiftvoid PhaseCFG::GlobalCodeMotion( Matcher &matcher, uint unique, Node_List &proj_list ) {
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift tty->print("\n---- Start GlobalCodeMotion ----\n");
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // Initialize the bbs.map for things on the proj_list
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // Set the basic block for Nodes pinned into blocks
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift Arena *a = Thread::current()->resource_area();
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // Find the earliest Block any instruction can be placed in. Some
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // instructions are pinned into Blocks. Unpinned instructions can
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // appear in last block in which all their inputs occur.
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift stack.map( (unique >> 1) + 16, NULL); // Pre-grow the list
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // Bailout without retry
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift C->record_method_not_compilable("early schedule failed");
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // Build Def-Use edges.
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift proj_list.push(_root); // Add real root as another root
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // Compute the latency information (via backwards walk) for all the
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // instructions in the graph
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift _node_latency = new GrowableArray<uint>(); // resource_area allocation
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // Now schedule all codes as LATE as possible. This is the LCA in the
6268e3d1b165bc72bbab62862e9b3b30e8883c97matthew_swift // dominator tree of all USES of a value. Pick the block with the least
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // loop nesting depth that is lowest in the dominator tree.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // ( visited.Clear() called in schedule_late()->Node_Backward_Iterator() )
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // schedule_late fails only when graph is incorrect.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(!VerifyGraphEdges, "verification should have failed");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print("\n---- Detect implicit null checks ----\n");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Detect implicit-null-check opportunities. Basically, find NULL checks
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // with suitable memory ops nearby. Use the memory op to do the NULL check.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // I can generate a memory op if there is not one nearby.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Don't do it for natives, adapters, or runtime stubs
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // ...and don't do it when there have been too many traps, globally.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (int reason = (int)Deoptimization::Reason_none+1;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(reason < BitsPerInt, "recode bit map");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (!C->too_many_traps((Deoptimization::DeoptReason) reason))
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // By reversing the loop direction we get a very minor gain on mpegaudio.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Feel free to revert to a forward loop for clarity.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // for( int i=0; i < (int)matcher._null_check_tests.size(); i+=2 ) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for( int i= matcher._null_check_tests.size()-2; i>=0; i-=2 ) {
6df5dcb0909ce8efb1aefa7a237a705e14de1f0bdugan _bbs[proj->_idx]->implicit_null_check(this, proj, val, allowed_reasons);
6df5dcb0909ce8efb1aefa7a237a705e14de1f0bdugan // The implicit_null_check will only perform the transformation
6df5dcb0909ce8efb1aefa7a237a705e14de1f0bdugan // if the null branch is truly uncommon, *and* it leads to an
6df5dcb0909ce8efb1aefa7a237a705e14de1f0bdugan // uncommon trap. Combined with the too_many_traps guards
6df5dcb0909ce8efb1aefa7a237a705e14de1f0bdugan // above, this prevents SEGV storms reported in 6366351,
6df5dcb0909ce8efb1aefa7a237a705e14de1f0bdugan // by recompiling offending methods without this optimization.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print("\n---- Start Local Scheduling ----\n");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Schedule locally. Right now a simple topological sort.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Later, do a real latency aware scheduler.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift int *ready_cnt = NEW_RESOURCE_ARRAY(int,C->unique());
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift memset( ready_cnt, -1, C->unique() * sizeof(int) );
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (i = 0; i < _num_blocks; i++) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (!_blocks[i]->schedule_local(this, matcher, ready_cnt, visited)) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (!C->failure_reason_is(C2Compiler::retry_no_subsuming_loads())) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift C->record_method_not_compilable("local schedule failed");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // If we inserted any instructions between a Call and his CatchNode,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // clone the instructions on all paths below the Catch.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for( i=0; i < _num_blocks; i++ )
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print("\n---- After GlobalCodeMotion ----\n");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift _node_latency = (GrowableArray<uint> *)0xdeadbeef;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------Estimate_Block_Frequency-----------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Estimate block frequencies based on IfNode probabilities.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Force conditional branches leading to uncommon traps to be unlikely,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // not because we get to the uncommon_trap with less relative frequency,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // but because an uncommon_trap typically causes a deopt, so we only get
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // there once.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (uint i = 1; i < root_blk->num_preds(); i++) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Create the loop tree and calculate loop depth.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Compute block frequency of each block, relative to a single loop entry.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Adjust all frequencies to be relative to a single method entry
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Save outmost loop frequency for LRG frequency threshold
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift _outer_loop_freq = _root_loop->outer_loop_freq();
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // force paths ending at uncommon traps to be infrequent
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (uint i = 1; i < root_blk->num_preds(); i++) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (pb->_num_succs == 1 && pb->_freq > PROB_MIN) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(b->_freq >= MIN_BLOCK_FREQUENCY, "Register Allocator requires meaningful block frequency");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//----------------------------create_loop_tree--------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Create a loop tree from the CFG
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Check that _loop field are clear...we could clear them if not.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(b->_loop == NULL, "clear _loop expected");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Sanity check that the RPO numbering is reflected in the _blocks array.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // It doesn't have to be for the loop tree to be built, but if it is not,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // then the blocks have been reordered since dom graph building...which
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // may question the RPO numbering
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(b->_rpo == i, "unexpected reverse post order number");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Assign blocks to loops
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for(uint i = _num_blocks - 1; i > 0; i-- ) { // skip Root block
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(loop_head->num_preds() - 1 == 2, "loop must have 2 predecessors");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift Node* tail_n = loop_head->pred(LoopNode::LoopBackControl);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Defensively filter out Loop nodes for non-single-entry loops.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // For all reasonable loops, the head occurs before the tail in RPO.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // The tail and (recursive) predecessors of the tail
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // are made members of a new loop.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(worklist.size() == 0, "nonempty worklist");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(loop_head->_loop == NULL, "just checking");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Add to nloop so push_pred() will skip over inner loops
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift nloop->push_pred(loop_head, LoopNode::LoopBackControl, worklist, _bbs);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (uint j = 1; j < member->num_preds(); j++) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Create a member list for each loop consisting
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // of both blocks and (immediate child) loops.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Not assigned to a loop. Add it to the method's pseudo loop.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (lp == root_loop || b != lp->head()) { // loop heads are already members
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Not a nested loop. Make it a child of the method's pseudo loop.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Add nested loop to member list of parent loop.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------push_pred--------------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftvoid CFGLoop::push_pred(Block* blk, int i, Block_List& worklist, Block_Array& node_to_blk) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Filter out blocks for non-single-entry loops.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // For all reasonable loops, the head occurs before the tail in RPO.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift } else if (pred_loop != this) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Nested loop.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift while (pred_loop->_parent != NULL && pred_loop->_parent != this) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Make pred's loop be a child
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Continue with loop entry predecessor.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(pred_head->num_preds() - 1 == 2, "loop must have 2 predecessors");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(pred_head != head(), "loop head in only one loop");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift push_pred(pred_head, LoopNode::EntryControl, worklist, node_to_blk);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(pred_loop->_parent == this && _parent == NULL, "just checking");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------add_nested_loop--------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Make cl a child of the current loop in the loop tree.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift while (ch->_sibling != NULL) { ch = ch->_sibling; }
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------compute_loop_depth-----------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Store the loop depth in each CFGLoop object.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Recursively walk the children to do the same for them.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------compute_freq-----------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Compute the frequency of each block and loop, relative to a single entry
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// into the dominating loop head.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Bottom up traversal of loop tree (visit inner loops first.)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Set loop head frequency to 1.0, then transitively
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // compute frequency for all successors in the loop,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // as well as for each exit edge. Inner loops are
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // treated as single blocks with loop exit targets
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // as the successor blocks.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Nested loops first
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert (_members.length() > 0, "no empty loops");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(lp->_parent == this, "immediate child");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (int k = 0; k < lp->_exits.length(); k++) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // For all loops other than the outer, "method" loop,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // sum and normalize the exit probability. The "method" loop
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // should keep the initial exit probability of 1, so that
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // inner blocks do not get erroneously scaled.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Total the exit probabilities for this loop.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Normalize the exit probabilities. Until now, the
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // probabilities estimate the possibility of exit per
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // a single loop iteration; afterward, they estimate
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // the probability of exit per loop entry.
b9a443e043c3d5c58c76dac2891abe1b71afa9d5matthew_swift new_prob = _exits.at(i).get_prob() / exits_sum;
b9a443e043c3d5c58c76dac2891abe1b71afa9d5matthew_swift // Save the total, but guard against unreasonable probability,
b9a443e043c3d5c58c76dac2891abe1b71afa9d5matthew_swift // as the value is used to estimate the loop trip count.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // An infinite trip count would blur relative block
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // frequencies.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (exits_sum < PROB_MIN) exits_sum = PROB_MIN;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------succ_prob-------------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Determine the probability of reaching successor 'i' from the receiver block.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Can only reach here if called after lcm. The original Op_If is gone,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // so we attempt to infer the probability from one or both of the
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // successor blocks.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(_num_succs == 2, "expecting 2 successors of a null check");
eb327f48ab993685cc144c4855d17a1fae0fdef5neil_a_wilson // If either successor has only one predecessor, then the
eb327f48ab993685cc144c4855d17a1fae0fdef5neil_a_wilson // probability estimate can be derived using the
eb327f48ab993685cc144c4855d17a1fae0fdef5neil_a_wilson // relative frequency of the successor and this block.
eb327f48ab993685cc144c4855d17a1fae0fdef5neil_a_wilson // Estimate using both successor frequencies
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Switch on branch type
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Conditionals pass on only part of their frequency
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(prob >= 0.0 && prob <= 1.0, "out of range probability");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // If succ[i] is the FALSE branch, invert path info
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if( _nodes[i + eidx + 1]->Opcode() == Op_IfFalse ) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Divide the frequency between all successors evenly
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift const CatchProjNode *ci = _nodes[i + eidx + 1]->as_CatchProj();
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (ci->_con == CatchProjNode::fall_through_index) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Fall-thru path gets the lion's share.
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift return 1.0f - PROB_UNLIKELY_MAG(5)*_num_succs;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Presume exceptional paths are equally unlikely
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift // Pass frequency straight thru to target
e32b0f1464ea290b749a43d5ee7a6085b14532c9matthew_swift // Do not push out freq to root block
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------num_fall_throughs-----------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Return the number of fall-through candidates for a block
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // In theory, either side can fall-thru, for simplicity sake,
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // let's say only the false branch can now.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Switch on branch type
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift const CatchProjNode *ci = _nodes[i + eidx + 1]->as_CatchProj();
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (ci->_con == CatchProjNode::fall_through_index) {
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift//------------------------------succ_fall_through-----------------------------
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift// Return true if a specific successor could be fall-through target.
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift // In theory, either side can fall-thru, for simplicity sake,
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift // let's say only the false branch can now.
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift return _nodes[i + eidx + 1]->Opcode() == Op_IfFalse;
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift // Switch on branch type
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift const CatchProjNode *ci = _nodes[i + eidx + 1]->as_CatchProj();
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift return ci->_con == CatchProjNode::fall_through_index;
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift return false;
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift return false;
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift//------------------------------update_uncommon_branch------------------------
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swift// Update the probability of a two-branch to be uncommon
408b8eb145348cb8f251a8abfec96e7e4e80f630matthew_swiftvoid Block::update_uncommon_branch(Block* ub) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(op == Op_CountedLoopEnd || op == Op_If, "must be a If");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(num_fall_throughs() == 2, "must be a two way branch block");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Which successor is ub?
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (s = 0; s <_num_succs; s++) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift assert(s < 2, "uncommon successor must be found");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // If ub is the true path, make the proability small, else
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // ub is the false path, and make the probability large
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift bool invert = (_nodes[s + eidx + 1]->Opcode() == Op_IfFalse);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Get existing probability
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------update_succ_freq-------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Update the appropriate frequency associated with block 'b', a successor of
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// a block in this loop.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swiftvoid CFGLoop::update_succ_freq(Block* b, float freq) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (b->_loop == this) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (b == head()) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // back branch within the loop
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // Do nothing now, the loop carried frequency will be
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // adjust later in scale_freq().
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // simple branch within the loop
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift } else if (!in_loop_nest(b)) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // branch is exit from this loop
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift // branch into nested loop
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------in_loop_nest-----------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Determine if block b is in the receiver's loop nest.
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift return b_loop == this;
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift//------------------------------scale_freq-------------------------------------
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Scale frequency of loops and blocks by trip counts from outer loops
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Do a top down traversal of loop tree (visit outer loops first.)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (g_isnan(block_freq) || block_freq < MIN_BLOCK_FREQUENCY)
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift// Frequency of outer loop
9355c1819517672f3cfe09356402fe0257c96facmatthew_swift//------------------------------dump_tree--------------------------------------
9355c1819517672f3cfe09356402fe0257c96facmatthew_swift//------------------------------dump-------------------------------------------
9355c1819517672f3cfe09356402fe0257c96facmatthew_swift for (int i = 0; i < _depth; i++) tty->print(" ");
9355c1819517672f3cfe09356402fe0257c96facmatthew_swift tty->print("%s: %d trip_count: %6.0f freq: %6.0f\n",
9355c1819517672f3cfe09356402fe0257c96facmatthew_swift _depth == 0 ? "Method" : "Loop", _id, trip_count(), _freq);
9355c1819517672f3cfe09356402fe0257c96facmatthew_swift for (int i = 0; i < _depth; i++) tty->print(" ");
9355c1819517672f3cfe09356402fe0257c96facmatthew_swift if (k++ >= 6) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (int j = 0; j < _depth+1; j++) tty->print(" ");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print(" B%d(%6.3f)", b->_pre_order, b->_freq);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print(" L%d(%6.3f)", lp->_id, lp->_freq);
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (int i = 0; i < _depth; i++) tty->print(" ");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift if (k++ >= 7) {
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift for (int j = 0; j < _depth+1; j++) tty->print(" ");
bb8874d71cdd8e5288297b9727703437c6dfcfedmatthew_swift tty->print(" ->%d@%d%%", blk->_pre_order, (int)(prob*100));