1 /*
2 * Copyright (c) 1997, 2017, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25 #include "precompiled.hpp"
26 #include "memory/allocation.inline.hpp"
27 #include "memory/resourceArea.hpp"
28 #include "opto/callnode.hpp"
29 #include "opto/chaitin.hpp"
30 #include "opto/live.hpp"
31 #include "opto/machnode.hpp"
32
33
34 // Compute live-in/live-out. We use a totally incremental algorithm. The LIVE
35 // problem is monotonic. The steady-state solution looks like this: pull a
36 // block from the worklist. It has a set of delta's - values which are newly
37 // live-in from the block. Push these to the live-out sets of all predecessor
38 // blocks. At each predecessor, the new live-out values are ANDed with what is
39 // already live-out (extra stuff is added to the live-out sets). Then the
40 // remaining new live-out values are ANDed with what is locally defined.
41 // Leftover bits become the new live-in for the predecessor block, and the pred
42 // block is put on the worklist.
43 // The locally live-in stuff is computed once and added to predecessor
44 // live-out sets. This separate compilation is done in the outer loop below.
PhaseLive(const PhaseCFG & cfg,const LRG_List & names,Arena * arena,bool keep_deltas)45 PhaseLive::PhaseLive(const PhaseCFG &cfg, const LRG_List &names, Arena *arena, bool keep_deltas)
46 : Phase(LIVE),
47 _live(0),
48 _livein(0),
49 _cfg(cfg),
50 _names(names),
51 _arena(arena),
52 _keep_deltas(keep_deltas) {
53 }
54
compute(uint maxlrg)55 void PhaseLive::compute(uint maxlrg) {
56 _maxlrg = maxlrg;
57 _worklist = new (_arena) Block_List();
58
59 // Init the sparse live arrays. This data is live on exit from here!
60 // The _live info is the live-out info.
61 _live = (IndexSet*)_arena->Amalloc(sizeof(IndexSet) * _cfg.number_of_blocks());
62 uint i;
63 for (i = 0; i < _cfg.number_of_blocks(); i++) {
64 _live[i].initialize(_maxlrg);
65 }
66
67 if (_keep_deltas) {
68 _livein = (IndexSet*)_arena->Amalloc(sizeof(IndexSet) * _cfg.number_of_blocks());
69 for (i = 0; i < _cfg.number_of_blocks(); i++) {
70 _livein[i].initialize(_maxlrg);
71 }
72 }
73
74 // Init the sparse arrays for delta-sets.
75 ResourceMark rm; // Nuke temp storage on exit
76
77 // Does the memory used by _defs and _deltas get reclaimed? Does it matter? TT
78
79 // Array of values defined locally in blocks
80 _defs = NEW_RESOURCE_ARRAY(IndexSet,_cfg.number_of_blocks());
81 for (i = 0; i < _cfg.number_of_blocks(); i++) {
82 _defs[i].initialize(_maxlrg);
83 }
84
85 // Array of delta-set pointers, indexed by block pre_order-1.
86 _deltas = NEW_RESOURCE_ARRAY(IndexSet*,_cfg.number_of_blocks());
87 memset(_deltas, 0, sizeof(IndexSet*)* _cfg.number_of_blocks());
88
89 _free_IndexSet = NULL;
90
91 // Blocks having done pass-1
92 VectorSet first_pass(Thread::current()->resource_area());
93
94 // Outer loop: must compute local live-in sets and push into predecessors.
95 for (uint j = _cfg.number_of_blocks(); j > 0; j--) {
96 Block* block = _cfg.get_block(j - 1);
97
98 // Compute the local live-in set. Start with any new live-out bits.
99 IndexSet* use = getset(block);
100 IndexSet* def = &_defs[block->_pre_order-1];
101 DEBUG_ONLY(IndexSet *def_outside = getfreeset();)
102 uint i;
103 for (i = block->number_of_nodes(); i > 1; i--) {
104 Node* n = block->get_node(i-1);
105 if (n->is_Phi()) {
106 break;
107 }
108
109 uint r = _names.at(n->_idx);
110 assert(!def_outside->member(r), "Use of external LRG overlaps the same LRG defined in this block");
111 def->insert(r);
112 use->remove(r);
113 uint cnt = n->req();
114 for (uint k = 1; k < cnt; k++) {
115 Node *nk = n->in(k);
116 uint nkidx = nk->_idx;
117 if (_cfg.get_block_for_node(nk) != block) {
118 uint u = _names.at(nkidx);
119 use->insert(u);
120 DEBUG_ONLY(def_outside->insert(u);)
121 }
122 }
123 }
124 #ifdef ASSERT
125 def_outside->set_next(_free_IndexSet);
126 _free_IndexSet = def_outside; // Drop onto free list
127 #endif
128 // Remove anything defined by Phis and the block start instruction
129 for (uint k = i; k > 0; k--) {
130 uint r = _names.at(block->get_node(k - 1)->_idx);
131 def->insert(r);
132 use->remove(r);
133 }
134
135 // Push these live-in things to predecessors
136 for (uint l = 1; l < block->num_preds(); l++) {
137 Block* p = _cfg.get_block_for_node(block->pred(l));
138 add_liveout(p, use, first_pass);
139
140 // PhiNode uses go in the live-out set of prior blocks.
141 for (uint k = i; k > 0; k--) {
142 Node *phi = block->get_node(k - 1);
143 if (l < phi->req()) {
144 add_liveout(p, _names.at(phi->in(l)->_idx), first_pass);
145 }
146 }
147 }
148 freeset(block);
149 first_pass.set(block->_pre_order);
150
151 // Inner loop: blocks that picked up new live-out values to be propagated
152 while (_worklist->size()) {
153 Block* block = _worklist->pop();
154 IndexSet *delta = getset(block);
155 assert(delta->count(), "missing delta set");
156
157 // Add new-live-in to predecessors live-out sets
158 for (uint l = 1; l < block->num_preds(); l++) {
159 Block* predecessor = _cfg.get_block_for_node(block->pred(l));
160 add_liveout(predecessor, delta, first_pass);
161 }
162
163 freeset(block);
164 } // End of while-worklist-not-empty
165
166 } // End of for-all-blocks-outer-loop
167
168 // We explicitly clear all of the IndexSets which we are about to release.
169 // This allows us to recycle their internal memory into IndexSet's free list.
170
171 for (i = 0; i < _cfg.number_of_blocks(); i++) {
172 _defs[i].clear();
173 if (_deltas[i]) {
174 // Is this always true?
175 _deltas[i]->clear();
176 }
177 }
178 IndexSet *free = _free_IndexSet;
179 while (free != NULL) {
180 IndexSet *temp = free;
181 free = free->next();
182 temp->clear();
183 }
184
185 }
186
187 #ifndef PRODUCT
stats(uint iters) const188 void PhaseLive::stats(uint iters) const {
189 }
190 #endif
191
192 // Get an IndexSet for a block. Return existing one, if any. Make a new
193 // empty one if a prior one does not exist.
getset(Block * p)194 IndexSet *PhaseLive::getset(Block *p) {
195 IndexSet *delta = _deltas[p->_pre_order-1];
196 if (!delta) { // Not on worklist?
197 // Get a free set; flag as being on worklist
198 delta = _deltas[p->_pre_order-1] = getfreeset();
199 }
200 return delta; // Return set of new live-out items
201 }
202
203 // Pull from free list, or allocate. Internal allocation on the returned set
204 // is always from thread local storage.
getfreeset()205 IndexSet *PhaseLive::getfreeset() {
206 IndexSet *f = _free_IndexSet;
207 if (!f) {
208 f = new IndexSet;
209 f->initialize(_maxlrg, Thread::current()->resource_area());
210 } else {
211 // Pull from free list
212 _free_IndexSet = f->next();
213 f->initialize(_maxlrg, Thread::current()->resource_area());
214 }
215 return f;
216 }
217
218 // Free an IndexSet from a block.
freeset(Block * p)219 void PhaseLive::freeset(Block *p) {
220 IndexSet *f = _deltas[p->_pre_order-1];
221 if (_keep_deltas) {
222 add_livein(p, f);
223 }
224 f->set_next(_free_IndexSet);
225 _free_IndexSet = f; // Drop onto free list
226 _deltas[p->_pre_order-1] = NULL;
227 }
228
229 // Add a live-out value to a given blocks live-out set. If it is new, then
230 // also add it to the delta set and stick the block on the worklist.
add_liveout(Block * p,uint r,VectorSet & first_pass)231 void PhaseLive::add_liveout(Block *p, uint r, VectorSet &first_pass) {
232 IndexSet *live = &_live[p->_pre_order-1];
233 if (live->insert(r)) { // If actually inserted...
234 // We extended the live-out set. See if the value is generated locally.
235 // If it is not, then we must extend the live-in set.
236 if (!_defs[p->_pre_order-1].member(r)) {
237 if (!_deltas[p->_pre_order-1] && // Not on worklist?
238 first_pass.test(p->_pre_order)) {
239 _worklist->push(p); // Actually go on worklist if already 1st pass
240 }
241 getset(p)->insert(r);
242 }
243 }
244 }
245
246 // Add a vector of live-out values to a given blocks live-out set.
add_liveout(Block * p,IndexSet * lo,VectorSet & first_pass)247 void PhaseLive::add_liveout(Block *p, IndexSet *lo, VectorSet &first_pass) {
248 IndexSet *live = &_live[p->_pre_order-1];
249 IndexSet *defs = &_defs[p->_pre_order-1];
250 IndexSet *on_worklist = _deltas[p->_pre_order-1];
251 IndexSet *delta = on_worklist ? on_worklist : getfreeset();
252
253 if (!lo->is_empty()) {
254 IndexSetIterator elements(lo);
255 uint r;
256 while ((r = elements.next()) != 0) {
257 if (live->insert(r) && // If actually inserted...
258 !defs->member(r)) { // and not defined locally
259 delta->insert(r); // Then add to live-in set
260 }
261 }
262 }
263
264 if (delta->count()) { // If actually added things
265 _deltas[p->_pre_order-1] = delta; // Flag as on worklist now
266 if (!on_worklist && // Not on worklist?
267 first_pass.test(p->_pre_order)) {
268 _worklist->push(p); // Actually go on worklist if already 1st pass
269 }
270 } else { // Nothing there; just free it
271 delta->set_next(_free_IndexSet);
272 _free_IndexSet = delta; // Drop onto free list
273 }
274 }
275
276 // Add a vector of live-in values to a given blocks live-in set.
add_livein(Block * p,IndexSet * lo)277 void PhaseLive::add_livein(Block *p, IndexSet *lo) {
278 IndexSet *livein = &_livein[p->_pre_order-1];
279 if (!livein->is_empty()) {
280 IndexSetIterator elements(lo);
281 uint r;
282 while ((r = elements.next()) != 0) {
283 livein->insert(r); // Then add to live-in set
284 }
285 }
286 }
287
288 #ifndef PRODUCT
289 // Dump the live-out set for a block
dump(const Block * b) const290 void PhaseLive::dump(const Block *b) const {
291 tty->print("Block %d: ",b->_pre_order);
292 if (_keep_deltas) {
293 tty->print("LiveIn: "); _livein[b->_pre_order-1].dump();
294 }
295 tty->print("LiveOut: "); _live[b->_pre_order-1].dump();
296 uint cnt = b->number_of_nodes();
297 for (uint i = 0; i < cnt; i++) {
298 tty->print("L%d/", _names.at(b->get_node(i)->_idx));
299 b->get_node(i)->dump();
300 }
301 tty->print("\n");
302 }
303
304 // Verify that base pointers and derived pointers are still sane.
verify_base_ptrs(ResourceArea * a) const305 void PhaseChaitin::verify_base_ptrs(ResourceArea *a) const {
306 #ifdef ASSERT
307 Unique_Node_List worklist(a);
308 for (uint i = 0; i < _cfg.number_of_blocks(); i++) {
309 Block* block = _cfg.get_block(i);
310 for (uint j = block->end_idx() + 1; j > 1; j--) {
311 Node* n = block->get_node(j-1);
312 if (n->is_Phi()) {
313 break;
314 }
315 // Found a safepoint?
316 if (n->is_MachSafePoint()) {
317 MachSafePointNode *sfpt = n->as_MachSafePoint();
318 JVMState* jvms = sfpt->jvms();
319 if (jvms != NULL) {
320 // Now scan for a live derived pointer
321 if (jvms->oopoff() < sfpt->req()) {
322 // Check each derived/base pair
323 for (uint idx = jvms->oopoff(); idx < sfpt->req(); idx++) {
324 Node *check = sfpt->in(idx);
325 bool is_derived = ((idx - jvms->oopoff()) & 1) == 0;
326 // search upwards through spills and spill phis for AddP
327 worklist.clear();
328 worklist.push(check);
329 uint k = 0;
330 while (k < worklist.size()) {
331 check = worklist.at(k);
332 assert(check,"Bad base or derived pointer");
333 // See PhaseChaitin::find_base_for_derived() for all cases.
334 int isc = check->is_Copy();
335 if (isc) {
336 worklist.push(check->in(isc));
337 } else if (check->is_Phi()) {
338 for (uint m = 1; m < check->req(); m++)
339 worklist.push(check->in(m));
340 } else if (check->is_Con()) {
341 if (is_derived) {
342 // Derived is NULL+offset
343 assert(!is_derived || check->bottom_type()->is_ptr()->ptr() == TypePtr::Null,"Bad derived pointer");
344 } else {
345 assert(check->bottom_type()->is_ptr()->_offset == 0,"Bad base pointer");
346 // Base either ConP(NULL) or loadConP
347 if (check->is_Mach()) {
348 assert(check->as_Mach()->ideal_Opcode() == Op_ConP,"Bad base pointer");
349 } else {
350 assert(check->Opcode() == Op_ConP &&
351 check->bottom_type()->is_ptr()->ptr() == TypePtr::Null,"Bad base pointer");
352 }
353 }
354 } else if (check->bottom_type()->is_ptr()->_offset == 0) {
355 if (check->is_Proj() || (check->is_Mach() &&
356 (check->as_Mach()->ideal_Opcode() == Op_CreateEx ||
357 check->as_Mach()->ideal_Opcode() == Op_ThreadLocal ||
358 check->as_Mach()->ideal_Opcode() == Op_CMoveP ||
359 check->as_Mach()->ideal_Opcode() == Op_CheckCastPP ||
360 #ifdef _LP64
361 (UseCompressedOops && check->as_Mach()->ideal_Opcode() == Op_CastPP) ||
362 (UseCompressedOops && check->as_Mach()->ideal_Opcode() == Op_DecodeN) ||
363 (UseCompressedClassPointers && check->as_Mach()->ideal_Opcode() == Op_DecodeNKlass) ||
364 #endif
365 check->as_Mach()->ideal_Opcode() == Op_LoadP ||
366 check->as_Mach()->ideal_Opcode() == Op_LoadKlass))) {
367 // Valid nodes
368 } else {
369 check->dump();
370 assert(false,"Bad base or derived pointer");
371 }
372 } else {
373 assert(is_derived,"Bad base pointer");
374 assert(check->is_Mach() && check->as_Mach()->ideal_Opcode() == Op_AddP,"Bad derived pointer");
375 }
376 k++;
377 assert(k < 100000,"Derived pointer checking in infinite loop");
378 } // End while
379 }
380 } // End of check for derived pointers
381 } // End of Kcheck for debug info
382 } // End of if found a safepoint
383 } // End of forall instructions in block
384 } // End of forall blocks
385 #endif
386 }
387
388 // Verify that graphs and base pointers are still sane.
verify(ResourceArea * a,bool verify_ifg) const389 void PhaseChaitin::verify(ResourceArea *a, bool verify_ifg) const {
390 #ifdef ASSERT
391 if (VerifyRegisterAllocator) {
392 _cfg.verify();
393 verify_base_ptrs(a);
394 if(verify_ifg)
395 _ifg->verify(this);
396 }
397 #endif
398 }
399
400 #endif
401