1/* Inlining decision heuristics.
2   Copyright (C) 2003-2015 Free Software Foundation, Inc.
3   Contributed by Jan Hubicka
4
5This file is part of GCC.
6
7GCC is free software; you can redistribute it and/or modify it under
8the terms of the GNU General Public License as published by the Free
9Software Foundation; either version 3, or (at your option) any later
10version.
11
12GCC is distributed in the hope that it will be useful, but WITHOUT ANY
13WARRANTY; without even the implied warranty of MERCHANTABILITY or
14FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
15for more details.
16
17You should have received a copy of the GNU General Public License
18along with GCC; see the file COPYING3.  If not see
19<http://www.gnu.org/licenses/>.  */
20
21/* Analysis used by the inliner and other passes limiting code size growth.
22
23   We estimate for each function
24     - function body size
25     - average function execution time
26     - inlining size benefit (that is how much of function body size
27       and its call sequence is expected to disappear by inlining)
28     - inlining time benefit
29     - function frame size
30   For each call
31     - call statement size and time
32
33   inlinie_summary datastructures store above information locally (i.e.
34   parameters of the function itself) and globally (i.e. parameters of
35   the function created by applying all the inline decisions already
36   present in the callgraph).
37
38   We provide accestor to the inline_summary datastructure and
39   basic logic updating the parameters when inlining is performed.
40
41   The summaries are context sensitive.  Context means
42     1) partial assignment of known constant values of operands
43     2) whether function is inlined into the call or not.
44   It is easy to add more variants.  To represent function size and time
45   that depends on context (i.e. it is known to be optimized away when
46   context is known either by inlining or from IP-CP and clonning),
47   we use predicates. Predicates are logical formulas in
48   conjunctive-disjunctive form consisting of clauses. Clauses are bitmaps
49   specifying what conditions must be true. Conditions are simple test
50   of the form described above.
51
52   In order to make predicate (possibly) true, all of its clauses must
53   be (possibly) true. To make clause (possibly) true, one of conditions
54   it mentions must be (possibly) true.  There are fixed bounds on
55   number of clauses and conditions and all the manipulation functions
56   are conservative in positive direction. I.e. we may lose precision
57   by thinking that predicate may be true even when it is not.
58
59   estimate_edge_size and estimate_edge_growth can be used to query
60   function size/time in the given context.  inline_merge_summary merges
61   properties of caller and callee after inlining.
62
63   Finally pass_inline_parameters is exported.  This is used to drive
64   computation of function parameters used by the early inliner. IPA
65   inlined performs analysis via its analyze_function method. */
66
67#include "config.h"
68#include "system.h"
69#include "coretypes.h"
70#include "tm.h"
71#include "hash-set.h"
72#include "machmode.h"
73#include "vec.h"
74#include "double-int.h"
75#include "input.h"
76#include "alias.h"
77#include "symtab.h"
78#include "wide-int.h"
79#include "inchash.h"
80#include "real.h"
81#include "tree.h"
82#include "fold-const.h"
83#include "stor-layout.h"
84#include "stringpool.h"
85#include "print-tree.h"
86#include "tree-inline.h"
87#include "langhooks.h"
88#include "flags.h"
89#include "diagnostic.h"
90#include "gimple-pretty-print.h"
91#include "params.h"
92#include "tree-pass.h"
93#include "coverage.h"
94#include "predict.h"
95#include "hard-reg-set.h"
96#include "input.h"
97#include "function.h"
98#include "dominance.h"
99#include "cfg.h"
100#include "cfganal.h"
101#include "basic-block.h"
102#include "tree-ssa-alias.h"
103#include "internal-fn.h"
104#include "gimple-expr.h"
105#include "is-a.h"
106#include "gimple.h"
107#include "gimple-iterator.h"
108#include "gimple-ssa.h"
109#include "tree-cfg.h"
110#include "tree-phinodes.h"
111#include "ssa-iterators.h"
112#include "tree-ssanames.h"
113#include "tree-ssa-loop-niter.h"
114#include "tree-ssa-loop.h"
115#include "hash-map.h"
116#include "plugin-api.h"
117#include "ipa-ref.h"
118#include "cgraph.h"
119#include "alloc-pool.h"
120#include "symbol-summary.h"
121#include "ipa-prop.h"
122#include "lto-streamer.h"
123#include "data-streamer.h"
124#include "tree-streamer.h"
125#include "ipa-inline.h"
126#include "cfgloop.h"
127#include "tree-scalar-evolution.h"
128#include "ipa-utils.h"
129#include "cilk.h"
130#include "cfgexpand.h"
131
132/* Estimate runtime of function can easilly run into huge numbers with many
133   nested loops.  Be sure we can compute time * INLINE_SIZE_SCALE * 2 in an
134   integer.  For anything larger we use gcov_type.  */
135#define MAX_TIME 500000
136
137/* Number of bits in integer, but we really want to be stable across different
138   hosts.  */
139#define NUM_CONDITIONS 32
140
141enum predicate_conditions
142{
143  predicate_false_condition = 0,
144  predicate_not_inlined_condition = 1,
145  predicate_first_dynamic_condition = 2
146};
147
148/* Special condition code we use to represent test that operand is compile time
149   constant.  */
150#define IS_NOT_CONSTANT ERROR_MARK
151/* Special condition code we use to represent test that operand is not changed
152   across invocation of the function.  When operand IS_NOT_CONSTANT it is always
153   CHANGED, however i.e. loop invariants can be NOT_CHANGED given percentage
154   of executions even when they are not compile time constants.  */
155#define CHANGED IDENTIFIER_NODE
156
157/* Holders of ipa cgraph hooks: */
158static struct cgraph_2edge_hook_list *edge_duplication_hook_holder;
159static struct cgraph_edge_hook_list *edge_removal_hook_holder;
160static void inline_edge_removal_hook (struct cgraph_edge *, void *);
161static void inline_edge_duplication_hook (struct cgraph_edge *,
162					  struct cgraph_edge *, void *);
163
164/* VECtor holding inline summaries.
165   In GGC memory because conditions might point to constant trees.  */
166function_summary <inline_summary *> *inline_summaries;
167vec<inline_edge_summary_t> inline_edge_summary_vec;
168
169/* Cached node/edge growths.  */
170vec<edge_growth_cache_entry> edge_growth_cache;
171
172/* Edge predicates goes here.  */
173static alloc_pool edge_predicate_pool;
174
175/* Return true predicate (tautology).
176   We represent it by empty list of clauses.  */
177
178static inline struct predicate
179true_predicate (void)
180{
181  struct predicate p;
182  p.clause[0] = 0;
183  return p;
184}
185
186
187/* Return predicate testing single condition number COND.  */
188
189static inline struct predicate
190single_cond_predicate (int cond)
191{
192  struct predicate p;
193  p.clause[0] = 1 << cond;
194  p.clause[1] = 0;
195  return p;
196}
197
198
199/* Return false predicate.  First clause require false condition.  */
200
201static inline struct predicate
202false_predicate (void)
203{
204  return single_cond_predicate (predicate_false_condition);
205}
206
207
208/* Return true if P is (true).  */
209
210static inline bool
211true_predicate_p (struct predicate *p)
212{
213  return !p->clause[0];
214}
215
216
217/* Return true if P is (false).  */
218
219static inline bool
220false_predicate_p (struct predicate *p)
221{
222  if (p->clause[0] == (1 << predicate_false_condition))
223    {
224      gcc_checking_assert (!p->clause[1]
225			   && p->clause[0] == 1 << predicate_false_condition);
226      return true;
227    }
228  return false;
229}
230
231
232/* Return predicate that is set true when function is not inlined.  */
233
234static inline struct predicate
235not_inlined_predicate (void)
236{
237  return single_cond_predicate (predicate_not_inlined_condition);
238}
239
240/* Simple description of whether a memory load or a condition refers to a load
241   from an aggregate and if so, how and where from in the aggregate.
242   Individual fields have the same meaning like fields with the same name in
243   struct condition.  */
244
245struct agg_position_info
246{
247  HOST_WIDE_INT offset;
248  bool agg_contents;
249  bool by_ref;
250};
251
252/* Add condition to condition list SUMMARY. OPERAND_NUM, SIZE, CODE and VAL
253   correspond to fields of condition structure.  AGGPOS describes whether the
254   used operand is loaded from an aggregate and where in the aggregate it is.
255   It can be NULL, which means this not a load from an aggregate.  */
256
257static struct predicate
258add_condition (struct inline_summary *summary, int operand_num,
259	       HOST_WIDE_INT size, struct agg_position_info *aggpos,
260	       enum tree_code code, tree val)
261{
262  int i;
263  struct condition *c;
264  struct condition new_cond;
265  HOST_WIDE_INT offset;
266  bool agg_contents, by_ref;
267
268  if (aggpos)
269    {
270      offset = aggpos->offset;
271      agg_contents = aggpos->agg_contents;
272      by_ref = aggpos->by_ref;
273    }
274  else
275    {
276      offset = 0;
277      agg_contents = false;
278      by_ref = false;
279    }
280
281  gcc_checking_assert (operand_num >= 0);
282  for (i = 0; vec_safe_iterate (summary->conds, i, &c); i++)
283    {
284      if (c->operand_num == operand_num
285	  && c->size == size
286	  && c->code == code
287	  && c->val == val
288	  && c->agg_contents == agg_contents
289	  && (!agg_contents || (c->offset == offset && c->by_ref == by_ref)))
290	return single_cond_predicate (i + predicate_first_dynamic_condition);
291    }
292  /* Too many conditions.  Give up and return constant true.  */
293  if (i == NUM_CONDITIONS - predicate_first_dynamic_condition)
294    return true_predicate ();
295
296  new_cond.operand_num = operand_num;
297  new_cond.code = code;
298  new_cond.val = val;
299  new_cond.agg_contents = agg_contents;
300  new_cond.by_ref = by_ref;
301  new_cond.offset = offset;
302  new_cond.size = size;
303  vec_safe_push (summary->conds, new_cond);
304  return single_cond_predicate (i + predicate_first_dynamic_condition);
305}
306
307
308/* Add clause CLAUSE into the predicate P.  */
309
310static inline void
311add_clause (conditions conditions, struct predicate *p, clause_t clause)
312{
313  int i;
314  int i2;
315  int insert_here = -1;
316  int c1, c2;
317
318  /* True clause.  */
319  if (!clause)
320    return;
321
322  /* False clause makes the whole predicate false.  Kill the other variants.  */
323  if (clause == (1 << predicate_false_condition))
324    {
325      p->clause[0] = (1 << predicate_false_condition);
326      p->clause[1] = 0;
327      return;
328    }
329  if (false_predicate_p (p))
330    return;
331
332  /* No one should be silly enough to add false into nontrivial clauses.  */
333  gcc_checking_assert (!(clause & (1 << predicate_false_condition)));
334
335  /* Look where to insert the clause.  At the same time prune out
336     clauses of P that are implied by the new clause and thus
337     redundant.  */
338  for (i = 0, i2 = 0; i <= MAX_CLAUSES; i++)
339    {
340      p->clause[i2] = p->clause[i];
341
342      if (!p->clause[i])
343	break;
344
345      /* If p->clause[i] implies clause, there is nothing to add.  */
346      if ((p->clause[i] & clause) == p->clause[i])
347	{
348	  /* We had nothing to add, none of clauses should've become
349	     redundant.  */
350	  gcc_checking_assert (i == i2);
351	  return;
352	}
353
354      if (p->clause[i] < clause && insert_here < 0)
355	insert_here = i2;
356
357      /* If clause implies p->clause[i], then p->clause[i] becomes redundant.
358         Otherwise the p->clause[i] has to stay.  */
359      if ((p->clause[i] & clause) != clause)
360	i2++;
361    }
362
363  /* Look for clauses that are obviously true.  I.e.
364     op0 == 5 || op0 != 5.  */
365  for (c1 = predicate_first_dynamic_condition; c1 < NUM_CONDITIONS; c1++)
366    {
367      condition *cc1;
368      if (!(clause & (1 << c1)))
369	continue;
370      cc1 = &(*conditions)[c1 - predicate_first_dynamic_condition];
371      /* We have no way to represent !CHANGED and !IS_NOT_CONSTANT
372         and thus there is no point for looking for them.  */
373      if (cc1->code == CHANGED || cc1->code == IS_NOT_CONSTANT)
374	continue;
375      for (c2 = c1 + 1; c2 < NUM_CONDITIONS; c2++)
376	if (clause & (1 << c2))
377	  {
378	    condition *cc1 =
379	      &(*conditions)[c1 - predicate_first_dynamic_condition];
380	    condition *cc2 =
381	      &(*conditions)[c2 - predicate_first_dynamic_condition];
382	    if (cc1->operand_num == cc2->operand_num
383		&& cc1->val == cc2->val
384		&& cc2->code != IS_NOT_CONSTANT
385		&& cc2->code != CHANGED
386		&& cc1->code == invert_tree_comparison (cc2->code,
387							HONOR_NANS (cc1->val)))
388	      return;
389	  }
390    }
391
392
393  /* We run out of variants.  Be conservative in positive direction.  */
394  if (i2 == MAX_CLAUSES)
395    return;
396  /* Keep clauses in decreasing order. This makes equivalence testing easy.  */
397  p->clause[i2 + 1] = 0;
398  if (insert_here >= 0)
399    for (; i2 > insert_here; i2--)
400      p->clause[i2] = p->clause[i2 - 1];
401  else
402    insert_here = i2;
403  p->clause[insert_here] = clause;
404}
405
406
407/* Return P & P2.  */
408
409static struct predicate
410and_predicates (conditions conditions,
411		struct predicate *p, struct predicate *p2)
412{
413  struct predicate out = *p;
414  int i;
415
416  /* Avoid busy work.  */
417  if (false_predicate_p (p2) || true_predicate_p (p))
418    return *p2;
419  if (false_predicate_p (p) || true_predicate_p (p2))
420    return *p;
421
422  /* See how far predicates match.  */
423  for (i = 0; p->clause[i] && p->clause[i] == p2->clause[i]; i++)
424    {
425      gcc_checking_assert (i < MAX_CLAUSES);
426    }
427
428  /* Combine the predicates rest.  */
429  for (; p2->clause[i]; i++)
430    {
431      gcc_checking_assert (i < MAX_CLAUSES);
432      add_clause (conditions, &out, p2->clause[i]);
433    }
434  return out;
435}
436
437
438/* Return true if predicates are obviously equal.  */
439
440static inline bool
441predicates_equal_p (struct predicate *p, struct predicate *p2)
442{
443  int i;
444  for (i = 0; p->clause[i]; i++)
445    {
446      gcc_checking_assert (i < MAX_CLAUSES);
447      gcc_checking_assert (p->clause[i] > p->clause[i + 1]);
448      gcc_checking_assert (!p2->clause[i]
449			   || p2->clause[i] > p2->clause[i + 1]);
450      if (p->clause[i] != p2->clause[i])
451	return false;
452    }
453  return !p2->clause[i];
454}
455
456
457/* Return P | P2.  */
458
459static struct predicate
460or_predicates (conditions conditions,
461	       struct predicate *p, struct predicate *p2)
462{
463  struct predicate out = true_predicate ();
464  int i, j;
465
466  /* Avoid busy work.  */
467  if (false_predicate_p (p2) || true_predicate_p (p))
468    return *p;
469  if (false_predicate_p (p) || true_predicate_p (p2))
470    return *p2;
471  if (predicates_equal_p (p, p2))
472    return *p;
473
474  /* OK, combine the predicates.  */
475  for (i = 0; p->clause[i]; i++)
476    for (j = 0; p2->clause[j]; j++)
477      {
478	gcc_checking_assert (i < MAX_CLAUSES && j < MAX_CLAUSES);
479	add_clause (conditions, &out, p->clause[i] | p2->clause[j]);
480      }
481  return out;
482}
483
484
485/* Having partial truth assignment in POSSIBLE_TRUTHS, return false
486   if predicate P is known to be false.  */
487
488static bool
489evaluate_predicate (struct predicate *p, clause_t possible_truths)
490{
491  int i;
492
493  /* True remains true.  */
494  if (true_predicate_p (p))
495    return true;
496
497  gcc_assert (!(possible_truths & (1 << predicate_false_condition)));
498
499  /* See if we can find clause we can disprove.  */
500  for (i = 0; p->clause[i]; i++)
501    {
502      gcc_checking_assert (i < MAX_CLAUSES);
503      if (!(p->clause[i] & possible_truths))
504	return false;
505    }
506  return true;
507}
508
509/* Return the probability in range 0...REG_BR_PROB_BASE that the predicated
510   instruction will be recomputed per invocation of the inlined call.  */
511
512static int
513predicate_probability (conditions conds,
514		       struct predicate *p, clause_t possible_truths,
515		       vec<inline_param_summary> inline_param_summary)
516{
517  int i;
518  int combined_prob = REG_BR_PROB_BASE;
519
520  /* True remains true.  */
521  if (true_predicate_p (p))
522    return REG_BR_PROB_BASE;
523
524  if (false_predicate_p (p))
525    return 0;
526
527  gcc_assert (!(possible_truths & (1 << predicate_false_condition)));
528
529  /* See if we can find clause we can disprove.  */
530  for (i = 0; p->clause[i]; i++)
531    {
532      gcc_checking_assert (i < MAX_CLAUSES);
533      if (!(p->clause[i] & possible_truths))
534	return 0;
535      else
536	{
537	  int this_prob = 0;
538	  int i2;
539	  if (!inline_param_summary.exists ())
540	    return REG_BR_PROB_BASE;
541	  for (i2 = 0; i2 < NUM_CONDITIONS; i2++)
542	    if ((p->clause[i] & possible_truths) & (1 << i2))
543	      {
544		if (i2 >= predicate_first_dynamic_condition)
545		  {
546		    condition *c =
547		      &(*conds)[i2 - predicate_first_dynamic_condition];
548		    if (c->code == CHANGED
549			&& (c->operand_num <
550			    (int) inline_param_summary.length ()))
551		      {
552			int iprob =
553			  inline_param_summary[c->operand_num].change_prob;
554			this_prob = MAX (this_prob, iprob);
555		      }
556		    else
557		      this_prob = REG_BR_PROB_BASE;
558		  }
559		else
560		  this_prob = REG_BR_PROB_BASE;
561	      }
562	  combined_prob = MIN (this_prob, combined_prob);
563	  if (!combined_prob)
564	    return 0;
565	}
566    }
567  return combined_prob;
568}
569
570
571/* Dump conditional COND.  */
572
573static void
574dump_condition (FILE *f, conditions conditions, int cond)
575{
576  condition *c;
577  if (cond == predicate_false_condition)
578    fprintf (f, "false");
579  else if (cond == predicate_not_inlined_condition)
580    fprintf (f, "not inlined");
581  else
582    {
583      c = &(*conditions)[cond - predicate_first_dynamic_condition];
584      fprintf (f, "op%i", c->operand_num);
585      if (c->agg_contents)
586	fprintf (f, "[%soffset: " HOST_WIDE_INT_PRINT_DEC "]",
587		 c->by_ref ? "ref " : "", c->offset);
588      if (c->code == IS_NOT_CONSTANT)
589	{
590	  fprintf (f, " not constant");
591	  return;
592	}
593      if (c->code == CHANGED)
594	{
595	  fprintf (f, " changed");
596	  return;
597	}
598      fprintf (f, " %s ", op_symbol_code (c->code));
599      print_generic_expr (f, c->val, 1);
600    }
601}
602
603
604/* Dump clause CLAUSE.  */
605
606static void
607dump_clause (FILE *f, conditions conds, clause_t clause)
608{
609  int i;
610  bool found = false;
611  fprintf (f, "(");
612  if (!clause)
613    fprintf (f, "true");
614  for (i = 0; i < NUM_CONDITIONS; i++)
615    if (clause & (1 << i))
616      {
617	if (found)
618	  fprintf (f, " || ");
619	found = true;
620	dump_condition (f, conds, i);
621      }
622  fprintf (f, ")");
623}
624
625
626/* Dump predicate PREDICATE.  */
627
628static void
629dump_predicate (FILE *f, conditions conds, struct predicate *pred)
630{
631  int i;
632  if (true_predicate_p (pred))
633    dump_clause (f, conds, 0);
634  else
635    for (i = 0; pred->clause[i]; i++)
636      {
637	if (i)
638	  fprintf (f, " && ");
639	dump_clause (f, conds, pred->clause[i]);
640      }
641  fprintf (f, "\n");
642}
643
644
645/* Dump inline hints.  */
646void
647dump_inline_hints (FILE *f, inline_hints hints)
648{
649  if (!hints)
650    return;
651  fprintf (f, "inline hints:");
652  if (hints & INLINE_HINT_indirect_call)
653    {
654      hints &= ~INLINE_HINT_indirect_call;
655      fprintf (f, " indirect_call");
656    }
657  if (hints & INLINE_HINT_loop_iterations)
658    {
659      hints &= ~INLINE_HINT_loop_iterations;
660      fprintf (f, " loop_iterations");
661    }
662  if (hints & INLINE_HINT_loop_stride)
663    {
664      hints &= ~INLINE_HINT_loop_stride;
665      fprintf (f, " loop_stride");
666    }
667  if (hints & INLINE_HINT_same_scc)
668    {
669      hints &= ~INLINE_HINT_same_scc;
670      fprintf (f, " same_scc");
671    }
672  if (hints & INLINE_HINT_in_scc)
673    {
674      hints &= ~INLINE_HINT_in_scc;
675      fprintf (f, " in_scc");
676    }
677  if (hints & INLINE_HINT_cross_module)
678    {
679      hints &= ~INLINE_HINT_cross_module;
680      fprintf (f, " cross_module");
681    }
682  if (hints & INLINE_HINT_declared_inline)
683    {
684      hints &= ~INLINE_HINT_declared_inline;
685      fprintf (f, " declared_inline");
686    }
687  if (hints & INLINE_HINT_array_index)
688    {
689      hints &= ~INLINE_HINT_array_index;
690      fprintf (f, " array_index");
691    }
692  if (hints & INLINE_HINT_known_hot)
693    {
694      hints &= ~INLINE_HINT_known_hot;
695      fprintf (f, " known_hot");
696    }
697  gcc_assert (!hints);
698}
699
700
701/* Record SIZE and TIME under condition PRED into the inline summary.  */
702
703static void
704account_size_time (struct inline_summary *summary, int size, int time,
705		   struct predicate *pred)
706{
707  size_time_entry *e;
708  bool found = false;
709  int i;
710
711  if (false_predicate_p (pred))
712    return;
713
714  /* We need to create initial empty unconitional clause, but otherwie
715     we don't need to account empty times and sizes.  */
716  if (!size && !time && summary->entry)
717    return;
718
719  /* Watch overflow that might result from insane profiles.  */
720  if (time > MAX_TIME * INLINE_TIME_SCALE)
721    time = MAX_TIME * INLINE_TIME_SCALE;
722  gcc_assert (time >= 0);
723
724  for (i = 0; vec_safe_iterate (summary->entry, i, &e); i++)
725    if (predicates_equal_p (&e->predicate, pred))
726      {
727	found = true;
728	break;
729      }
730  if (i == 256)
731    {
732      i = 0;
733      found = true;
734      e = &(*summary->entry)[0];
735      gcc_assert (!e->predicate.clause[0]);
736      if (dump_file && (dump_flags & TDF_DETAILS))
737	fprintf (dump_file,
738		 "\t\tReached limit on number of entries, "
739		 "ignoring the predicate.");
740    }
741  if (dump_file && (dump_flags & TDF_DETAILS) && (time || size))
742    {
743      fprintf (dump_file,
744	       "\t\tAccounting size:%3.2f, time:%3.2f on %spredicate:",
745	       ((double) size) / INLINE_SIZE_SCALE,
746	       ((double) time) / INLINE_TIME_SCALE, found ? "" : "new ");
747      dump_predicate (dump_file, summary->conds, pred);
748    }
749  if (!found)
750    {
751      struct size_time_entry new_entry;
752      new_entry.size = size;
753      new_entry.time = time;
754      new_entry.predicate = *pred;
755      vec_safe_push (summary->entry, new_entry);
756    }
757  else
758    {
759      e->size += size;
760      e->time += time;
761      if (e->time > MAX_TIME * INLINE_TIME_SCALE)
762	e->time = MAX_TIME * INLINE_TIME_SCALE;
763    }
764}
765
766/* We proved E to be unreachable, redirect it to __bultin_unreachable.  */
767
768static struct cgraph_edge *
769redirect_to_unreachable (struct cgraph_edge *e)
770{
771  struct cgraph_node *callee = !e->inline_failed ? e->callee : NULL;
772  struct cgraph_node *target = cgraph_node::get_create
773		      (builtin_decl_implicit (BUILT_IN_UNREACHABLE));
774
775  if (e->speculative)
776    e = e->resolve_speculation (target->decl);
777  else if (!e->callee)
778    e->make_direct (target);
779  else
780    e->redirect_callee (target);
781  struct inline_edge_summary *es = inline_edge_summary (e);
782  e->inline_failed = CIF_UNREACHABLE;
783  e->frequency = 0;
784  e->count = 0;
785  es->call_stmt_size = 0;
786  es->call_stmt_time = 0;
787  if (callee)
788    callee->remove_symbol_and_inline_clones ();
789  return e;
790}
791
792/* Set predicate for edge E.  */
793
794static void
795edge_set_predicate (struct cgraph_edge *e, struct predicate *predicate)
796{
797  /* If the edge is determined to be never executed, redirect it
798     to BUILTIN_UNREACHABLE to save inliner from inlining into it.  */
799  if (predicate && false_predicate_p (predicate)
800      /* When handling speculative edges, we need to do the redirection
801         just once.  Do it always on the direct edge, so we do not
802	 attempt to resolve speculation while duplicating the edge.  */
803      && (!e->speculative || e->callee))
804    e = redirect_to_unreachable (e);
805
806  struct inline_edge_summary *es = inline_edge_summary (e);
807  if (predicate && !true_predicate_p (predicate))
808    {
809      if (!es->predicate)
810	es->predicate = (struct predicate *) pool_alloc (edge_predicate_pool);
811      *es->predicate = *predicate;
812    }
813  else
814    {
815      if (es->predicate)
816	pool_free (edge_predicate_pool, es->predicate);
817      es->predicate = NULL;
818    }
819}
820
821/* Set predicate for hint *P.  */
822
823static void
824set_hint_predicate (struct predicate **p, struct predicate new_predicate)
825{
826  if (false_predicate_p (&new_predicate) || true_predicate_p (&new_predicate))
827    {
828      if (*p)
829	pool_free (edge_predicate_pool, *p);
830      *p = NULL;
831    }
832  else
833    {
834      if (!*p)
835	*p = (struct predicate *) pool_alloc (edge_predicate_pool);
836      **p = new_predicate;
837    }
838}
839
840
841/* KNOWN_VALS is partial mapping of parameters of NODE to constant values.
842   KNOWN_AGGS is a vector of aggreggate jump functions for each parameter.
843   Return clause of possible truths. When INLINE_P is true, assume that we are
844   inlining.
845
846   ERROR_MARK means compile time invariant.  */
847
848static clause_t
849evaluate_conditions_for_known_args (struct cgraph_node *node,
850				    bool inline_p,
851				    vec<tree> known_vals,
852				    vec<ipa_agg_jump_function_p>
853				    known_aggs)
854{
855  clause_t clause = inline_p ? 0 : 1 << predicate_not_inlined_condition;
856  struct inline_summary *info = inline_summaries->get (node);
857  int i;
858  struct condition *c;
859
860  for (i = 0; vec_safe_iterate (info->conds, i, &c); i++)
861    {
862      tree val;
863      tree res;
864
865      /* We allow call stmt to have fewer arguments than the callee function
866         (especially for K&R style programs).  So bound check here (we assume
867         known_aggs vector, if non-NULL, has the same length as
868         known_vals).  */
869      gcc_checking_assert (!known_aggs.exists ()
870			   || (known_vals.length () == known_aggs.length ()));
871      if (c->operand_num >= (int) known_vals.length ())
872	{
873	  clause |= 1 << (i + predicate_first_dynamic_condition);
874	  continue;
875	}
876
877      if (c->agg_contents)
878	{
879	  struct ipa_agg_jump_function *agg;
880
881	  if (c->code == CHANGED
882	      && !c->by_ref
883	      && (known_vals[c->operand_num] == error_mark_node))
884	    continue;
885
886	  if (known_aggs.exists ())
887	    {
888	      agg = known_aggs[c->operand_num];
889	      val = ipa_find_agg_cst_for_param (agg, c->offset, c->by_ref);
890	    }
891	  else
892	    val = NULL_TREE;
893	}
894      else
895	{
896	  val = known_vals[c->operand_num];
897	  if (val == error_mark_node && c->code != CHANGED)
898	    val = NULL_TREE;
899	}
900
901      if (!val)
902	{
903	  clause |= 1 << (i + predicate_first_dynamic_condition);
904	  continue;
905	}
906      if (c->code == CHANGED)
907	continue;
908
909      if (tree_to_shwi (TYPE_SIZE (TREE_TYPE (val))) != c->size)
910	{
911	  clause |= 1 << (i + predicate_first_dynamic_condition);
912	  continue;
913	}
914      if (c->code == IS_NOT_CONSTANT)
915	continue;
916
917      val = fold_unary (VIEW_CONVERT_EXPR, TREE_TYPE (c->val), val);
918      res = val
919	? fold_binary_to_constant (c->code, boolean_type_node, val, c->val)
920	: NULL;
921
922      if (res && integer_zerop (res))
923	continue;
924
925      clause |= 1 << (i + predicate_first_dynamic_condition);
926    }
927  return clause;
928}
929
930
931/* Work out what conditions might be true at invocation of E.  */
932
933static void
934evaluate_properties_for_edge (struct cgraph_edge *e, bool inline_p,
935			      clause_t *clause_ptr,
936			      vec<tree> *known_vals_ptr,
937			      vec<ipa_polymorphic_call_context>
938			      *known_contexts_ptr,
939			      vec<ipa_agg_jump_function_p> *known_aggs_ptr)
940{
941  struct cgraph_node *callee = e->callee->ultimate_alias_target ();
942  struct inline_summary *info = inline_summaries->get (callee);
943  vec<tree> known_vals = vNULL;
944  vec<ipa_agg_jump_function_p> known_aggs = vNULL;
945
946  if (clause_ptr)
947    *clause_ptr = inline_p ? 0 : 1 << predicate_not_inlined_condition;
948  if (known_vals_ptr)
949    known_vals_ptr->create (0);
950  if (known_contexts_ptr)
951    known_contexts_ptr->create (0);
952
953  if (ipa_node_params_sum
954      && !e->call_stmt_cannot_inline_p
955      && ((clause_ptr && info->conds) || known_vals_ptr || known_contexts_ptr))
956    {
957      struct ipa_node_params *parms_info;
958      struct ipa_edge_args *args = IPA_EDGE_REF (e);
959      struct inline_edge_summary *es = inline_edge_summary (e);
960      int i, count = ipa_get_cs_argument_count (args);
961
962      if (e->caller->global.inlined_to)
963	parms_info = IPA_NODE_REF (e->caller->global.inlined_to);
964      else
965	parms_info = IPA_NODE_REF (e->caller);
966
967      if (count && (info->conds || known_vals_ptr))
968	known_vals.safe_grow_cleared (count);
969      if (count && (info->conds || known_aggs_ptr))
970	known_aggs.safe_grow_cleared (count);
971      if (count && known_contexts_ptr)
972	known_contexts_ptr->safe_grow_cleared (count);
973
974      for (i = 0; i < count; i++)
975	{
976	  struct ipa_jump_func *jf = ipa_get_ith_jump_func (args, i);
977	  tree cst = ipa_value_from_jfunc (parms_info, jf);
978
979	  if (!cst && e->call_stmt
980	      && i < (int)gimple_call_num_args (e->call_stmt))
981	    {
982	      cst = gimple_call_arg (e->call_stmt, i);
983	      if (!is_gimple_min_invariant (cst))
984		cst = NULL;
985	    }
986	  if (cst)
987	    {
988	      gcc_checking_assert (TREE_CODE (cst) != TREE_BINFO);
989	      if (known_vals.exists ())
990		known_vals[i] = cst;
991	    }
992	  else if (inline_p && !es->param[i].change_prob)
993	    known_vals[i] = error_mark_node;
994
995	  if (known_contexts_ptr)
996	    (*known_contexts_ptr)[i] = ipa_context_from_jfunc (parms_info, e,
997							       i, jf);
998	  /* TODO: When IPA-CP starts propagating and merging aggregate jump
999	     functions, use its knowledge of the caller too, just like the
1000	     scalar case above.  */
1001	  known_aggs[i] = &jf->agg;
1002	}
1003    }
1004  else if (e->call_stmt && !e->call_stmt_cannot_inline_p
1005	   && ((clause_ptr && info->conds) || known_vals_ptr))
1006    {
1007      int i, count = (int)gimple_call_num_args (e->call_stmt);
1008
1009      if (count && (info->conds || known_vals_ptr))
1010	known_vals.safe_grow_cleared (count);
1011      for (i = 0; i < count; i++)
1012	{
1013	  tree cst = gimple_call_arg (e->call_stmt, i);
1014	  if (!is_gimple_min_invariant (cst))
1015	    cst = NULL;
1016	  if (cst)
1017	    known_vals[i] = cst;
1018	}
1019    }
1020
1021  if (clause_ptr)
1022    *clause_ptr = evaluate_conditions_for_known_args (callee, inline_p,
1023						      known_vals, known_aggs);
1024
1025  if (known_vals_ptr)
1026    *known_vals_ptr = known_vals;
1027  else
1028    known_vals.release ();
1029
1030  if (known_aggs_ptr)
1031    *known_aggs_ptr = known_aggs;
1032  else
1033    known_aggs.release ();
1034}
1035
1036
1037/* Allocate the inline summary vector or resize it to cover all cgraph nodes. */
1038
1039static void
1040inline_summary_alloc (void)
1041{
1042  if (!edge_removal_hook_holder)
1043    edge_removal_hook_holder =
1044      symtab->add_edge_removal_hook (&inline_edge_removal_hook, NULL);
1045  if (!edge_duplication_hook_holder)
1046    edge_duplication_hook_holder =
1047      symtab->add_edge_duplication_hook (&inline_edge_duplication_hook, NULL);
1048
1049  if (!inline_summaries)
1050    inline_summaries = (inline_summary_t*) inline_summary_t::create_ggc (symtab);
1051
1052  if (inline_edge_summary_vec.length () <= (unsigned) symtab->edges_max_uid)
1053    inline_edge_summary_vec.safe_grow_cleared (symtab->edges_max_uid + 1);
1054  if (!edge_predicate_pool)
1055    edge_predicate_pool = create_alloc_pool ("edge predicates",
1056					     sizeof (struct predicate), 10);
1057}
1058
1059/* We are called multiple time for given function; clear
1060   data from previous run so they are not cumulated.  */
1061
1062static void
1063reset_inline_edge_summary (struct cgraph_edge *e)
1064{
1065  if (e->uid < (int) inline_edge_summary_vec.length ())
1066    {
1067      struct inline_edge_summary *es = inline_edge_summary (e);
1068
1069      es->call_stmt_size = es->call_stmt_time = 0;
1070      if (es->predicate)
1071	pool_free (edge_predicate_pool, es->predicate);
1072      es->predicate = NULL;
1073      es->param.release ();
1074    }
1075}
1076
1077/* We are called multiple time for given function; clear
1078   data from previous run so they are not cumulated.  */
1079
1080static void
1081reset_inline_summary (struct cgraph_node *node,
1082		      inline_summary *info)
1083{
1084  struct cgraph_edge *e;
1085
1086  info->self_size = info->self_time = 0;
1087  info->estimated_stack_size = 0;
1088  info->estimated_self_stack_size = 0;
1089  info->stack_frame_offset = 0;
1090  info->size = 0;
1091  info->time = 0;
1092  info->growth = 0;
1093  info->scc_no = 0;
1094  if (info->loop_iterations)
1095    {
1096      pool_free (edge_predicate_pool, info->loop_iterations);
1097      info->loop_iterations = NULL;
1098    }
1099  if (info->loop_stride)
1100    {
1101      pool_free (edge_predicate_pool, info->loop_stride);
1102      info->loop_stride = NULL;
1103    }
1104  if (info->array_index)
1105    {
1106      pool_free (edge_predicate_pool, info->array_index);
1107      info->array_index = NULL;
1108    }
1109  vec_free (info->conds);
1110  vec_free (info->entry);
1111  for (e = node->callees; e; e = e->next_callee)
1112    reset_inline_edge_summary (e);
1113  for (e = node->indirect_calls; e; e = e->next_callee)
1114    reset_inline_edge_summary (e);
1115}
1116
1117/* Hook that is called by cgraph.c when a node is removed.  */
1118
1119void
1120inline_summary_t::remove (cgraph_node *node, inline_summary *info)
1121{
1122  reset_inline_summary (node, info);
1123}
1124
1125/* Remap predicate P of former function to be predicate of duplicated function.
1126   POSSIBLE_TRUTHS is clause of possible truths in the duplicated node,
1127   INFO is inline summary of the duplicated node.  */
1128
1129static struct predicate
1130remap_predicate_after_duplication (struct predicate *p,
1131				   clause_t possible_truths,
1132				   struct inline_summary *info)
1133{
1134  struct predicate new_predicate = true_predicate ();
1135  int j;
1136  for (j = 0; p->clause[j]; j++)
1137    if (!(possible_truths & p->clause[j]))
1138      {
1139	new_predicate = false_predicate ();
1140	break;
1141      }
1142    else
1143      add_clause (info->conds, &new_predicate,
1144		  possible_truths & p->clause[j]);
1145  return new_predicate;
1146}
1147
1148/* Same as remap_predicate_after_duplication but handle hint predicate *P.
1149   Additionally care about allocating new memory slot for updated predicate
1150   and set it to NULL when it becomes true or false (and thus uninteresting).
1151 */
1152
1153static void
1154remap_hint_predicate_after_duplication (struct predicate **p,
1155					clause_t possible_truths,
1156					struct inline_summary *info)
1157{
1158  struct predicate new_predicate;
1159
1160  if (!*p)
1161    return;
1162
1163  new_predicate = remap_predicate_after_duplication (*p,
1164						     possible_truths, info);
1165  /* We do not want to free previous predicate; it is used by node origin.  */
1166  *p = NULL;
1167  set_hint_predicate (p, new_predicate);
1168}
1169
1170
1171/* Hook that is called by cgraph.c when a node is duplicated.  */
1172void
1173inline_summary_t::duplicate (cgraph_node *src,
1174			     cgraph_node *dst,
1175			     inline_summary *,
1176			     inline_summary *info)
1177{
1178  inline_summary_alloc ();
1179  memcpy (info, inline_summaries->get (src), sizeof (inline_summary));
1180  /* TODO: as an optimization, we may avoid copying conditions
1181     that are known to be false or true.  */
1182  info->conds = vec_safe_copy (info->conds);
1183
1184  /* When there are any replacements in the function body, see if we can figure
1185     out that something was optimized out.  */
1186  if (ipa_node_params_sum && dst->clone.tree_map)
1187    {
1188      vec<size_time_entry, va_gc> *entry = info->entry;
1189      /* Use SRC parm info since it may not be copied yet.  */
1190      struct ipa_node_params *parms_info = IPA_NODE_REF (src);
1191      vec<tree> known_vals = vNULL;
1192      int count = ipa_get_param_count (parms_info);
1193      int i, j;
1194      clause_t possible_truths;
1195      struct predicate true_pred = true_predicate ();
1196      size_time_entry *e;
1197      int optimized_out_size = 0;
1198      bool inlined_to_p = false;
1199      struct cgraph_edge *edge, *next;
1200
1201      info->entry = 0;
1202      known_vals.safe_grow_cleared (count);
1203      for (i = 0; i < count; i++)
1204	{
1205	  struct ipa_replace_map *r;
1206
1207	  for (j = 0; vec_safe_iterate (dst->clone.tree_map, j, &r); j++)
1208	    {
1209	      if (((!r->old_tree && r->parm_num == i)
1210		   || (r->old_tree && r->old_tree == ipa_get_param (parms_info, i)))
1211		   && r->replace_p && !r->ref_p)
1212		{
1213		  known_vals[i] = r->new_tree;
1214		  break;
1215		}
1216	    }
1217	}
1218      possible_truths = evaluate_conditions_for_known_args (dst, false,
1219							    known_vals,
1220							    vNULL);
1221      known_vals.release ();
1222
1223      account_size_time (info, 0, 0, &true_pred);
1224
1225      /* Remap size_time vectors.
1226         Simplify the predicate by prunning out alternatives that are known
1227         to be false.
1228         TODO: as on optimization, we can also eliminate conditions known
1229         to be true.  */
1230      for (i = 0; vec_safe_iterate (entry, i, &e); i++)
1231	{
1232	  struct predicate new_predicate;
1233	  new_predicate = remap_predicate_after_duplication (&e->predicate,
1234							     possible_truths,
1235							     info);
1236	  if (false_predicate_p (&new_predicate))
1237	    optimized_out_size += e->size;
1238	  else
1239	    account_size_time (info, e->size, e->time, &new_predicate);
1240	}
1241
1242      /* Remap edge predicates with the same simplification as above.
1243         Also copy constantness arrays.   */
1244      for (edge = dst->callees; edge; edge = next)
1245	{
1246	  struct predicate new_predicate;
1247	  struct inline_edge_summary *es = inline_edge_summary (edge);
1248	  next = edge->next_callee;
1249
1250	  if (!edge->inline_failed)
1251	    inlined_to_p = true;
1252	  if (!es->predicate)
1253	    continue;
1254	  new_predicate = remap_predicate_after_duplication (es->predicate,
1255							     possible_truths,
1256							     info);
1257	  if (false_predicate_p (&new_predicate)
1258	      && !false_predicate_p (es->predicate))
1259	    optimized_out_size += es->call_stmt_size * INLINE_SIZE_SCALE;
1260	  edge_set_predicate (edge, &new_predicate);
1261	}
1262
1263      /* Remap indirect edge predicates with the same simplificaiton as above.
1264         Also copy constantness arrays.   */
1265      for (edge = dst->indirect_calls; edge; edge = next)
1266	{
1267	  struct predicate new_predicate;
1268	  struct inline_edge_summary *es = inline_edge_summary (edge);
1269	  next = edge->next_callee;
1270
1271	  gcc_checking_assert (edge->inline_failed);
1272	  if (!es->predicate)
1273	    continue;
1274	  new_predicate = remap_predicate_after_duplication (es->predicate,
1275							     possible_truths,
1276							     info);
1277	  if (false_predicate_p (&new_predicate)
1278	      && !false_predicate_p (es->predicate))
1279	    optimized_out_size += es->call_stmt_size * INLINE_SIZE_SCALE;
1280	  edge_set_predicate (edge, &new_predicate);
1281	}
1282      remap_hint_predicate_after_duplication (&info->loop_iterations,
1283					      possible_truths, info);
1284      remap_hint_predicate_after_duplication (&info->loop_stride,
1285					      possible_truths, info);
1286      remap_hint_predicate_after_duplication (&info->array_index,
1287					      possible_truths, info);
1288
1289      /* If inliner or someone after inliner will ever start producing
1290         non-trivial clones, we will get trouble with lack of information
1291         about updating self sizes, because size vectors already contains
1292         sizes of the calees.  */
1293      gcc_assert (!inlined_to_p || !optimized_out_size);
1294    }
1295  else
1296    {
1297      info->entry = vec_safe_copy (info->entry);
1298      if (info->loop_iterations)
1299	{
1300	  predicate p = *info->loop_iterations;
1301	  info->loop_iterations = NULL;
1302	  set_hint_predicate (&info->loop_iterations, p);
1303	}
1304      if (info->loop_stride)
1305	{
1306	  predicate p = *info->loop_stride;
1307	  info->loop_stride = NULL;
1308	  set_hint_predicate (&info->loop_stride, p);
1309	}
1310      if (info->array_index)
1311	{
1312	  predicate p = *info->array_index;
1313	  info->array_index = NULL;
1314	  set_hint_predicate (&info->array_index, p);
1315	}
1316    }
1317  if (!dst->global.inlined_to)
1318    inline_update_overall_summary (dst);
1319}
1320
1321
1322/* Hook that is called by cgraph.c when a node is duplicated.  */
1323
1324static void
1325inline_edge_duplication_hook (struct cgraph_edge *src,
1326			      struct cgraph_edge *dst,
1327			      ATTRIBUTE_UNUSED void *data)
1328{
1329  struct inline_edge_summary *info;
1330  struct inline_edge_summary *srcinfo;
1331  inline_summary_alloc ();
1332  info = inline_edge_summary (dst);
1333  srcinfo = inline_edge_summary (src);
1334  memcpy (info, srcinfo, sizeof (struct inline_edge_summary));
1335  info->predicate = NULL;
1336  edge_set_predicate (dst, srcinfo->predicate);
1337  info->param = srcinfo->param.copy ();
1338  if (!dst->indirect_unknown_callee && src->indirect_unknown_callee)
1339    {
1340      info->call_stmt_size -= (eni_size_weights.indirect_call_cost
1341			       - eni_size_weights.call_cost);
1342      info->call_stmt_time -= (eni_time_weights.indirect_call_cost
1343			       - eni_time_weights.call_cost);
1344    }
1345}
1346
1347
1348/* Keep edge cache consistent across edge removal.  */
1349
1350static void
1351inline_edge_removal_hook (struct cgraph_edge *edge,
1352			  void *data ATTRIBUTE_UNUSED)
1353{
1354  if (edge_growth_cache.exists ())
1355    reset_edge_growth_cache (edge);
1356  reset_inline_edge_summary (edge);
1357}
1358
1359
1360/* Initialize growth caches.  */
1361
1362void
1363initialize_growth_caches (void)
1364{
1365  if (symtab->edges_max_uid)
1366    edge_growth_cache.safe_grow_cleared (symtab->edges_max_uid);
1367}
1368
1369
1370/* Free growth caches.  */
1371
1372void
1373free_growth_caches (void)
1374{
1375  edge_growth_cache.release ();
1376}
1377
1378
1379/* Dump edge summaries associated to NODE and recursively to all clones.
1380   Indent by INDENT.  */
1381
1382static void
1383dump_inline_edge_summary (FILE *f, int indent, struct cgraph_node *node,
1384			  struct inline_summary *info)
1385{
1386  struct cgraph_edge *edge;
1387  for (edge = node->callees; edge; edge = edge->next_callee)
1388    {
1389      struct inline_edge_summary *es = inline_edge_summary (edge);
1390      struct cgraph_node *callee = edge->callee->ultimate_alias_target ();
1391      int i;
1392
1393      fprintf (f,
1394	       "%*s%s/%i %s\n%*s  loop depth:%2i freq:%4i size:%2i"
1395	       " time: %2i callee size:%2i stack:%2i",
1396	       indent, "", callee->name (), callee->order,
1397	       !edge->inline_failed
1398	       ? "inlined" : cgraph_inline_failed_string (edge-> inline_failed),
1399	       indent, "", es->loop_depth, edge->frequency,
1400	       es->call_stmt_size, es->call_stmt_time,
1401	       (int) inline_summaries->get (callee)->size / INLINE_SIZE_SCALE,
1402	       (int) inline_summaries->get (callee)->estimated_stack_size);
1403
1404      if (es->predicate)
1405	{
1406	  fprintf (f, " predicate: ");
1407	  dump_predicate (f, info->conds, es->predicate);
1408	}
1409      else
1410	fprintf (f, "\n");
1411      if (es->param.exists ())
1412	for (i = 0; i < (int) es->param.length (); i++)
1413	  {
1414	    int prob = es->param[i].change_prob;
1415
1416	    if (!prob)
1417	      fprintf (f, "%*s op%i is compile time invariant\n",
1418		       indent + 2, "", i);
1419	    else if (prob != REG_BR_PROB_BASE)
1420	      fprintf (f, "%*s op%i change %f%% of time\n", indent + 2, "", i,
1421		       prob * 100.0 / REG_BR_PROB_BASE);
1422	  }
1423      if (!edge->inline_failed)
1424	{
1425	  fprintf (f, "%*sStack frame offset %i, callee self size %i,"
1426		   " callee size %i\n",
1427		   indent + 2, "",
1428		   (int) inline_summaries->get (callee)->stack_frame_offset,
1429		   (int) inline_summaries->get (callee)->estimated_self_stack_size,
1430		   (int) inline_summaries->get (callee)->estimated_stack_size);
1431	  dump_inline_edge_summary (f, indent + 2, callee, info);
1432	}
1433    }
1434  for (edge = node->indirect_calls; edge; edge = edge->next_callee)
1435    {
1436      struct inline_edge_summary *es = inline_edge_summary (edge);
1437      fprintf (f, "%*sindirect call loop depth:%2i freq:%4i size:%2i"
1438	       " time: %2i",
1439	       indent, "",
1440	       es->loop_depth,
1441	       edge->frequency, es->call_stmt_size, es->call_stmt_time);
1442      if (es->predicate)
1443	{
1444	  fprintf (f, "predicate: ");
1445	  dump_predicate (f, info->conds, es->predicate);
1446	}
1447      else
1448	fprintf (f, "\n");
1449    }
1450}
1451
1452
1453void
1454dump_inline_summary (FILE *f, struct cgraph_node *node)
1455{
1456  if (node->definition)
1457    {
1458      struct inline_summary *s = inline_summaries->get (node);
1459      size_time_entry *e;
1460      int i;
1461      fprintf (f, "Inline summary for %s/%i", node->name (),
1462	       node->order);
1463      if (DECL_DISREGARD_INLINE_LIMITS (node->decl))
1464	fprintf (f, " always_inline");
1465      if (s->inlinable)
1466	fprintf (f, " inlinable");
1467      if (s->contains_cilk_spawn)
1468	fprintf (f, " contains_cilk_spawn");
1469      fprintf (f, "\n  self time:       %i\n", s->self_time);
1470      fprintf (f, "  global time:     %i\n", s->time);
1471      fprintf (f, "  self size:       %i\n", s->self_size);
1472      fprintf (f, "  global size:     %i\n", s->size);
1473      fprintf (f, "  min size:       %i\n", s->min_size);
1474      fprintf (f, "  self stack:      %i\n",
1475	       (int) s->estimated_self_stack_size);
1476      fprintf (f, "  global stack:    %i\n", (int) s->estimated_stack_size);
1477      if (s->growth)
1478	fprintf (f, "  estimated growth:%i\n", (int) s->growth);
1479      if (s->scc_no)
1480	fprintf (f, "  In SCC:          %i\n", (int) s->scc_no);
1481      for (i = 0; vec_safe_iterate (s->entry, i, &e); i++)
1482	{
1483	  fprintf (f, "    size:%f, time:%f, predicate:",
1484		   (double) e->size / INLINE_SIZE_SCALE,
1485		   (double) e->time / INLINE_TIME_SCALE);
1486	  dump_predicate (f, s->conds, &e->predicate);
1487	}
1488      if (s->loop_iterations)
1489	{
1490	  fprintf (f, "  loop iterations:");
1491	  dump_predicate (f, s->conds, s->loop_iterations);
1492	}
1493      if (s->loop_stride)
1494	{
1495	  fprintf (f, "  loop stride:");
1496	  dump_predicate (f, s->conds, s->loop_stride);
1497	}
1498      if (s->array_index)
1499	{
1500	  fprintf (f, "  array index:");
1501	  dump_predicate (f, s->conds, s->array_index);
1502	}
1503      fprintf (f, "  calls:\n");
1504      dump_inline_edge_summary (f, 4, node, s);
1505      fprintf (f, "\n");
1506    }
1507}
1508
1509DEBUG_FUNCTION void
1510debug_inline_summary (struct cgraph_node *node)
1511{
1512  dump_inline_summary (stderr, node);
1513}
1514
1515void
1516dump_inline_summaries (FILE *f)
1517{
1518  struct cgraph_node *node;
1519
1520  FOR_EACH_DEFINED_FUNCTION (node)
1521    if (!node->global.inlined_to)
1522      dump_inline_summary (f, node);
1523}
1524
1525/* Give initial reasons why inlining would fail on EDGE.  This gets either
1526   nullified or usually overwritten by more precise reasons later.  */
1527
1528void
1529initialize_inline_failed (struct cgraph_edge *e)
1530{
1531  struct cgraph_node *callee = e->callee;
1532
1533  if (e->indirect_unknown_callee)
1534    e->inline_failed = CIF_INDIRECT_UNKNOWN_CALL;
1535  else if (!callee->definition)
1536    e->inline_failed = CIF_BODY_NOT_AVAILABLE;
1537  else if (callee->local.redefined_extern_inline)
1538    e->inline_failed = CIF_REDEFINED_EXTERN_INLINE;
1539  else if (e->call_stmt_cannot_inline_p)
1540    e->inline_failed = CIF_MISMATCHED_ARGUMENTS;
1541  else if (cfun && fn_contains_cilk_spawn_p (cfun))
1542    /* We can't inline if the function is spawing a function.  */
1543    e->inline_failed = CIF_FUNCTION_NOT_INLINABLE;
1544  else
1545    e->inline_failed = CIF_FUNCTION_NOT_CONSIDERED;
1546}
1547
1548/* Callback of walk_aliased_vdefs.  Flags that it has been invoked to the
1549   boolean variable pointed to by DATA.  */
1550
1551static bool
1552mark_modified (ao_ref *ao ATTRIBUTE_UNUSED, tree vdef ATTRIBUTE_UNUSED,
1553	       void *data)
1554{
1555  bool *b = (bool *) data;
1556  *b = true;
1557  return true;
1558}
1559
1560/* If OP refers to value of function parameter, return the corresponding
1561   parameter.  If non-NULL, the size of the memory load (or the SSA_NAME of the
1562   PARM_DECL) will be stored to *SIZE_P in that case too.  */
1563
1564static tree
1565unmodified_parm_1 (gimple stmt, tree op, HOST_WIDE_INT *size_p)
1566{
1567  /* SSA_NAME referring to parm default def?  */
1568  if (TREE_CODE (op) == SSA_NAME
1569      && SSA_NAME_IS_DEFAULT_DEF (op)
1570      && TREE_CODE (SSA_NAME_VAR (op)) == PARM_DECL)
1571    {
1572      if (size_p)
1573	*size_p = tree_to_shwi (TYPE_SIZE (TREE_TYPE (op)));
1574      return SSA_NAME_VAR (op);
1575    }
1576  /* Non-SSA parm reference?  */
1577  if (TREE_CODE (op) == PARM_DECL)
1578    {
1579      bool modified = false;
1580
1581      ao_ref refd;
1582      ao_ref_init (&refd, op);
1583      walk_aliased_vdefs (&refd, gimple_vuse (stmt), mark_modified, &modified,
1584			  NULL);
1585      if (!modified)
1586	{
1587	  if (size_p)
1588	    *size_p = tree_to_shwi (TYPE_SIZE (TREE_TYPE (op)));
1589	  return op;
1590	}
1591    }
1592  return NULL_TREE;
1593}
1594
1595/* If OP refers to value of function parameter, return the corresponding
1596   parameter.  Also traverse chains of SSA register assignments.  If non-NULL,
1597   the size of the memory load (or the SSA_NAME of the PARM_DECL) will be
1598   stored to *SIZE_P in that case too.  */
1599
1600static tree
1601unmodified_parm (gimple stmt, tree op, HOST_WIDE_INT *size_p)
1602{
1603  tree res = unmodified_parm_1 (stmt, op, size_p);
1604  if (res)
1605    return res;
1606
1607  if (TREE_CODE (op) == SSA_NAME
1608      && !SSA_NAME_IS_DEFAULT_DEF (op)
1609      && gimple_assign_single_p (SSA_NAME_DEF_STMT (op)))
1610    return unmodified_parm (SSA_NAME_DEF_STMT (op),
1611			    gimple_assign_rhs1 (SSA_NAME_DEF_STMT (op)),
1612			    size_p);
1613  return NULL_TREE;
1614}
1615
1616/* If OP refers to a value of a function parameter or value loaded from an
1617   aggregate passed to a parameter (either by value or reference), return TRUE
1618   and store the number of the parameter to *INDEX_P, the access size into
1619   *SIZE_P, and information whether and how it has been loaded from an
1620   aggregate into *AGGPOS.  INFO describes the function parameters, STMT is the
1621   statement in which OP is used or loaded.  */
1622
1623static bool
1624unmodified_parm_or_parm_agg_item (struct ipa_func_body_info *fbi,
1625				  gimple stmt, tree op, int *index_p,
1626				  HOST_WIDE_INT *size_p,
1627				  struct agg_position_info *aggpos)
1628{
1629  tree res = unmodified_parm_1 (stmt, op, size_p);
1630
1631  gcc_checking_assert (aggpos);
1632  if (res)
1633    {
1634      *index_p = ipa_get_param_decl_index (fbi->info, res);
1635      if (*index_p < 0)
1636	return false;
1637      aggpos->agg_contents = false;
1638      aggpos->by_ref = false;
1639      return true;
1640    }
1641
1642  if (TREE_CODE (op) == SSA_NAME)
1643    {
1644      if (SSA_NAME_IS_DEFAULT_DEF (op)
1645	  || !gimple_assign_single_p (SSA_NAME_DEF_STMT (op)))
1646	return false;
1647      stmt = SSA_NAME_DEF_STMT (op);
1648      op = gimple_assign_rhs1 (stmt);
1649      if (!REFERENCE_CLASS_P (op))
1650	return unmodified_parm_or_parm_agg_item (fbi, stmt, op, index_p, size_p,
1651						 aggpos);
1652    }
1653
1654  aggpos->agg_contents = true;
1655  return ipa_load_from_parm_agg (fbi, fbi->info->descriptors,
1656				 stmt, op, index_p, &aggpos->offset,
1657				 size_p, &aggpos->by_ref);
1658}
1659
1660/* See if statement might disappear after inlining.
1661   0 - means not eliminated
1662   1 - half of statements goes away
1663   2 - for sure it is eliminated.
1664   We are not terribly sophisticated, basically looking for simple abstraction
1665   penalty wrappers.  */
1666
1667static int
1668eliminated_by_inlining_prob (gimple stmt)
1669{
1670  enum gimple_code code = gimple_code (stmt);
1671  enum tree_code rhs_code;
1672
1673  if (!optimize)
1674    return 0;
1675
1676  switch (code)
1677    {
1678    case GIMPLE_RETURN:
1679      return 2;
1680    case GIMPLE_ASSIGN:
1681      if (gimple_num_ops (stmt) != 2)
1682	return 0;
1683
1684      rhs_code = gimple_assign_rhs_code (stmt);
1685
1686      /* Casts of parameters, loads from parameters passed by reference
1687         and stores to return value or parameters are often free after
1688         inlining dua to SRA and further combining.
1689         Assume that half of statements goes away.  */
1690      if (CONVERT_EXPR_CODE_P (rhs_code)
1691	  || rhs_code == VIEW_CONVERT_EXPR
1692	  || rhs_code == ADDR_EXPR
1693	  || gimple_assign_rhs_class (stmt) == GIMPLE_SINGLE_RHS)
1694	{
1695	  tree rhs = gimple_assign_rhs1 (stmt);
1696	  tree lhs = gimple_assign_lhs (stmt);
1697	  tree inner_rhs = get_base_address (rhs);
1698	  tree inner_lhs = get_base_address (lhs);
1699	  bool rhs_free = false;
1700	  bool lhs_free = false;
1701
1702	  if (!inner_rhs)
1703	    inner_rhs = rhs;
1704	  if (!inner_lhs)
1705	    inner_lhs = lhs;
1706
1707	  /* Reads of parameter are expected to be free.  */
1708	  if (unmodified_parm (stmt, inner_rhs, NULL))
1709	    rhs_free = true;
1710	  /* Match expressions of form &this->field. Those will most likely
1711	     combine with something upstream after inlining.  */
1712	  else if (TREE_CODE (inner_rhs) == ADDR_EXPR)
1713	    {
1714	      tree op = get_base_address (TREE_OPERAND (inner_rhs, 0));
1715	      if (TREE_CODE (op) == PARM_DECL)
1716		rhs_free = true;
1717	      else if (TREE_CODE (op) == MEM_REF
1718		       && unmodified_parm (stmt, TREE_OPERAND (op, 0), NULL))
1719		rhs_free = true;
1720	    }
1721
1722	  /* When parameter is not SSA register because its address is taken
1723	     and it is just copied into one, the statement will be completely
1724	     free after inlining (we will copy propagate backward).   */
1725	  if (rhs_free && is_gimple_reg (lhs))
1726	    return 2;
1727
1728	  /* Reads of parameters passed by reference
1729	     expected to be free (i.e. optimized out after inlining).  */
1730	  if (TREE_CODE (inner_rhs) == MEM_REF
1731	      && unmodified_parm (stmt, TREE_OPERAND (inner_rhs, 0), NULL))
1732	    rhs_free = true;
1733
1734	  /* Copying parameter passed by reference into gimple register is
1735	     probably also going to copy propagate, but we can't be quite
1736	     sure.  */
1737	  if (rhs_free && is_gimple_reg (lhs))
1738	    lhs_free = true;
1739
1740	  /* Writes to parameters, parameters passed by value and return value
1741	     (either dirrectly or passed via invisible reference) are free.
1742
1743	     TODO: We ought to handle testcase like
1744	     struct a {int a,b;};
1745	     struct a
1746	     retrurnsturct (void)
1747	     {
1748	     struct a a ={1,2};
1749	     return a;
1750	     }
1751
1752	     This translate into:
1753
1754	     retrurnsturct ()
1755	     {
1756	     int a$b;
1757	     int a$a;
1758	     struct a a;
1759	     struct a D.2739;
1760
1761	     <bb 2>:
1762	     D.2739.a = 1;
1763	     D.2739.b = 2;
1764	     return D.2739;
1765
1766	     }
1767	     For that we either need to copy ipa-split logic detecting writes
1768	     to return value.  */
1769	  if (TREE_CODE (inner_lhs) == PARM_DECL
1770	      || TREE_CODE (inner_lhs) == RESULT_DECL
1771	      || (TREE_CODE (inner_lhs) == MEM_REF
1772		  && (unmodified_parm (stmt, TREE_OPERAND (inner_lhs, 0), NULL)
1773		      || (TREE_CODE (TREE_OPERAND (inner_lhs, 0)) == SSA_NAME
1774			  && SSA_NAME_VAR (TREE_OPERAND (inner_lhs, 0))
1775			  && TREE_CODE (SSA_NAME_VAR (TREE_OPERAND
1776						      (inner_lhs,
1777						       0))) == RESULT_DECL))))
1778	    lhs_free = true;
1779	  if (lhs_free
1780	      && (is_gimple_reg (rhs) || is_gimple_min_invariant (rhs)))
1781	    rhs_free = true;
1782	  if (lhs_free && rhs_free)
1783	    return 1;
1784	}
1785      return 0;
1786    default:
1787      return 0;
1788    }
1789}
1790
1791
1792/* If BB ends by a conditional we can turn into predicates, attach corresponding
1793   predicates to the CFG edges.   */
1794
1795static void
1796set_cond_stmt_execution_predicate (struct ipa_func_body_info *fbi,
1797				   struct inline_summary *summary,
1798				   basic_block bb)
1799{
1800  gimple last;
1801  tree op;
1802  HOST_WIDE_INT size;
1803  int index;
1804  struct agg_position_info aggpos;
1805  enum tree_code code, inverted_code;
1806  edge e;
1807  edge_iterator ei;
1808  gimple set_stmt;
1809  tree op2;
1810
1811  last = last_stmt (bb);
1812  if (!last || gimple_code (last) != GIMPLE_COND)
1813    return;
1814  if (!is_gimple_ip_invariant (gimple_cond_rhs (last)))
1815    return;
1816  op = gimple_cond_lhs (last);
1817  /* TODO: handle conditionals like
1818     var = op0 < 4;
1819     if (var != 0).  */
1820  if (unmodified_parm_or_parm_agg_item (fbi, last, op, &index, &size, &aggpos))
1821    {
1822      code = gimple_cond_code (last);
1823      inverted_code = invert_tree_comparison (code, HONOR_NANS (op));
1824
1825      FOR_EACH_EDGE (e, ei, bb->succs)
1826	{
1827	  enum tree_code this_code = (e->flags & EDGE_TRUE_VALUE
1828				      ? code : inverted_code);
1829	  /* invert_tree_comparison will return ERROR_MARK on FP
1830	     comparsions that are not EQ/NE instead of returning proper
1831	     unordered one.  Be sure it is not confused with NON_CONSTANT.  */
1832	  if (this_code != ERROR_MARK)
1833	    {
1834	      struct predicate p = add_condition (summary, index, size, &aggpos,
1835						  this_code,
1836						  gimple_cond_rhs (last));
1837	      e->aux = pool_alloc (edge_predicate_pool);
1838	      *(struct predicate *) e->aux = p;
1839	    }
1840	}
1841    }
1842
1843  if (TREE_CODE (op) != SSA_NAME)
1844    return;
1845  /* Special case
1846     if (builtin_constant_p (op))
1847     constant_code
1848     else
1849     nonconstant_code.
1850     Here we can predicate nonconstant_code.  We can't
1851     really handle constant_code since we have no predicate
1852     for this and also the constant code is not known to be
1853     optimized away when inliner doen't see operand is constant.
1854     Other optimizers might think otherwise.  */
1855  if (gimple_cond_code (last) != NE_EXPR
1856      || !integer_zerop (gimple_cond_rhs (last)))
1857    return;
1858  set_stmt = SSA_NAME_DEF_STMT (op);
1859  if (!gimple_call_builtin_p (set_stmt, BUILT_IN_CONSTANT_P)
1860      || gimple_call_num_args (set_stmt) != 1)
1861    return;
1862  op2 = gimple_call_arg (set_stmt, 0);
1863  if (!unmodified_parm_or_parm_agg_item (fbi, set_stmt, op2, &index, &size,
1864					 &aggpos))
1865    return;
1866  FOR_EACH_EDGE (e, ei, bb->succs) if (e->flags & EDGE_FALSE_VALUE)
1867    {
1868      struct predicate p = add_condition (summary, index, size, &aggpos,
1869					  IS_NOT_CONSTANT, NULL_TREE);
1870      e->aux = pool_alloc (edge_predicate_pool);
1871      *(struct predicate *) e->aux = p;
1872    }
1873}
1874
1875
1876/* If BB ends by a switch we can turn into predicates, attach corresponding
1877   predicates to the CFG edges.   */
1878
1879static void
1880set_switch_stmt_execution_predicate (struct ipa_func_body_info *fbi,
1881				     struct inline_summary *summary,
1882				     basic_block bb)
1883{
1884  gimple lastg;
1885  tree op;
1886  int index;
1887  HOST_WIDE_INT size;
1888  struct agg_position_info aggpos;
1889  edge e;
1890  edge_iterator ei;
1891  size_t n;
1892  size_t case_idx;
1893
1894  lastg = last_stmt (bb);
1895  if (!lastg || gimple_code (lastg) != GIMPLE_SWITCH)
1896    return;
1897  gswitch *last = as_a <gswitch *> (lastg);
1898  op = gimple_switch_index (last);
1899  if (!unmodified_parm_or_parm_agg_item (fbi, last, op, &index, &size, &aggpos))
1900    return;
1901
1902  FOR_EACH_EDGE (e, ei, bb->succs)
1903    {
1904      e->aux = pool_alloc (edge_predicate_pool);
1905      *(struct predicate *) e->aux = false_predicate ();
1906    }
1907  n = gimple_switch_num_labels (last);
1908  for (case_idx = 0; case_idx < n; ++case_idx)
1909    {
1910      tree cl = gimple_switch_label (last, case_idx);
1911      tree min, max;
1912      struct predicate p;
1913
1914      e = find_edge (bb, label_to_block (CASE_LABEL (cl)));
1915      min = CASE_LOW (cl);
1916      max = CASE_HIGH (cl);
1917
1918      /* For default we might want to construct predicate that none
1919         of cases is met, but it is bit hard to do not having negations
1920         of conditionals handy.  */
1921      if (!min && !max)
1922	p = true_predicate ();
1923      else if (!max)
1924	p = add_condition (summary, index, size, &aggpos, EQ_EXPR, min);
1925      else
1926	{
1927	  struct predicate p1, p2;
1928	  p1 = add_condition (summary, index, size, &aggpos, GE_EXPR, min);
1929	  p2 = add_condition (summary, index, size, &aggpos, LE_EXPR, max);
1930	  p = and_predicates (summary->conds, &p1, &p2);
1931	}
1932      *(struct predicate *) e->aux
1933	= or_predicates (summary->conds, &p, (struct predicate *) e->aux);
1934    }
1935}
1936
1937
1938/* For each BB in NODE attach to its AUX pointer predicate under
1939   which it is executable.  */
1940
1941static void
1942compute_bb_predicates (struct ipa_func_body_info *fbi,
1943		       struct cgraph_node *node,
1944		       struct inline_summary *summary)
1945{
1946  struct function *my_function = DECL_STRUCT_FUNCTION (node->decl);
1947  bool done = false;
1948  basic_block bb;
1949
1950  FOR_EACH_BB_FN (bb, my_function)
1951    {
1952      set_cond_stmt_execution_predicate (fbi, summary, bb);
1953      set_switch_stmt_execution_predicate (fbi, summary, bb);
1954    }
1955
1956  /* Entry block is always executable.  */
1957  ENTRY_BLOCK_PTR_FOR_FN (my_function)->aux
1958    = pool_alloc (edge_predicate_pool);
1959  *(struct predicate *) ENTRY_BLOCK_PTR_FOR_FN (my_function)->aux
1960    = true_predicate ();
1961
1962  /* A simple dataflow propagation of predicates forward in the CFG.
1963     TODO: work in reverse postorder.  */
1964  while (!done)
1965    {
1966      done = true;
1967      FOR_EACH_BB_FN (bb, my_function)
1968	{
1969	  struct predicate p = false_predicate ();
1970	  edge e;
1971	  edge_iterator ei;
1972	  FOR_EACH_EDGE (e, ei, bb->preds)
1973	    {
1974	      if (e->src->aux)
1975		{
1976		  struct predicate this_bb_predicate
1977		    = *(struct predicate *) e->src->aux;
1978		  if (e->aux)
1979		    this_bb_predicate
1980		      = and_predicates (summary->conds, &this_bb_predicate,
1981					(struct predicate *) e->aux);
1982		  p = or_predicates (summary->conds, &p, &this_bb_predicate);
1983		  if (true_predicate_p (&p))
1984		    break;
1985		}
1986	    }
1987	  if (false_predicate_p (&p))
1988	    gcc_assert (!bb->aux);
1989	  else
1990	    {
1991	      if (!bb->aux)
1992		{
1993		  done = false;
1994		  bb->aux = pool_alloc (edge_predicate_pool);
1995		  *((struct predicate *) bb->aux) = p;
1996		}
1997	      else if (!predicates_equal_p (&p, (struct predicate *) bb->aux))
1998		{
1999		  /* This OR operation is needed to ensure monotonous data flow
2000		     in the case we hit the limit on number of clauses and the
2001		     and/or operations above give approximate answers.  */
2002		  p = or_predicates (summary->conds, &p, (struct predicate *)bb->aux);
2003	          if (!predicates_equal_p (&p, (struct predicate *) bb->aux))
2004		    {
2005		      done = false;
2006		      *((struct predicate *) bb->aux) = p;
2007		    }
2008		}
2009	    }
2010	}
2011    }
2012}
2013
2014
2015/* We keep info about constantness of SSA names.  */
2016
2017typedef struct predicate predicate_t;
2018/* Return predicate specifying when the STMT might have result that is not
2019   a compile time constant.  */
2020
2021static struct predicate
2022will_be_nonconstant_expr_predicate (struct ipa_node_params *info,
2023				    struct inline_summary *summary,
2024				    tree expr,
2025				    vec<predicate_t> nonconstant_names)
2026{
2027  tree parm;
2028  int index;
2029  HOST_WIDE_INT size;
2030
2031  while (UNARY_CLASS_P (expr))
2032    expr = TREE_OPERAND (expr, 0);
2033
2034  parm = unmodified_parm (NULL, expr, &size);
2035  if (parm && (index = ipa_get_param_decl_index (info, parm)) >= 0)
2036    return add_condition (summary, index, size, NULL, CHANGED, NULL_TREE);
2037  if (is_gimple_min_invariant (expr))
2038    return false_predicate ();
2039  if (TREE_CODE (expr) == SSA_NAME)
2040    return nonconstant_names[SSA_NAME_VERSION (expr)];
2041  if (BINARY_CLASS_P (expr) || COMPARISON_CLASS_P (expr))
2042    {
2043      struct predicate p1 = will_be_nonconstant_expr_predicate
2044	(info, summary, TREE_OPERAND (expr, 0),
2045	 nonconstant_names);
2046      struct predicate p2;
2047      if (true_predicate_p (&p1))
2048	return p1;
2049      p2 = will_be_nonconstant_expr_predicate (info, summary,
2050					       TREE_OPERAND (expr, 1),
2051					       nonconstant_names);
2052      return or_predicates (summary->conds, &p1, &p2);
2053    }
2054  else if (TREE_CODE (expr) == COND_EXPR)
2055    {
2056      struct predicate p1 = will_be_nonconstant_expr_predicate
2057	(info, summary, TREE_OPERAND (expr, 0),
2058	 nonconstant_names);
2059      struct predicate p2;
2060      if (true_predicate_p (&p1))
2061	return p1;
2062      p2 = will_be_nonconstant_expr_predicate (info, summary,
2063					       TREE_OPERAND (expr, 1),
2064					       nonconstant_names);
2065      if (true_predicate_p (&p2))
2066	return p2;
2067      p1 = or_predicates (summary->conds, &p1, &p2);
2068      p2 = will_be_nonconstant_expr_predicate (info, summary,
2069					       TREE_OPERAND (expr, 2),
2070					       nonconstant_names);
2071      return or_predicates (summary->conds, &p1, &p2);
2072    }
2073  else
2074    {
2075      debug_tree (expr);
2076      gcc_unreachable ();
2077    }
2078  return false_predicate ();
2079}
2080
2081
2082/* Return predicate specifying when the STMT might have result that is not
2083   a compile time constant.  */
2084
2085static struct predicate
2086will_be_nonconstant_predicate (struct ipa_func_body_info *fbi,
2087			       struct inline_summary *summary,
2088			       gimple stmt,
2089			       vec<predicate_t> nonconstant_names)
2090{
2091  struct predicate p = true_predicate ();
2092  ssa_op_iter iter;
2093  tree use;
2094  struct predicate op_non_const;
2095  bool is_load;
2096  int base_index;
2097  HOST_WIDE_INT size;
2098  struct agg_position_info aggpos;
2099
2100  /* What statments might be optimized away
2101     when their arguments are constant.  */
2102  if (gimple_code (stmt) != GIMPLE_ASSIGN
2103      && gimple_code (stmt) != GIMPLE_COND
2104      && gimple_code (stmt) != GIMPLE_SWITCH
2105      && (gimple_code (stmt) != GIMPLE_CALL
2106	  || !(gimple_call_flags (stmt) & ECF_CONST)))
2107    return p;
2108
2109  /* Stores will stay anyway.  */
2110  if (gimple_store_p (stmt))
2111    return p;
2112
2113  is_load = gimple_assign_load_p (stmt);
2114
2115  /* Loads can be optimized when the value is known.  */
2116  if (is_load)
2117    {
2118      tree op;
2119      gcc_assert (gimple_assign_single_p (stmt));
2120      op = gimple_assign_rhs1 (stmt);
2121      if (!unmodified_parm_or_parm_agg_item (fbi, stmt, op, &base_index, &size,
2122					     &aggpos))
2123	return p;
2124    }
2125  else
2126    base_index = -1;
2127
2128  /* See if we understand all operands before we start
2129     adding conditionals.  */
2130  FOR_EACH_SSA_TREE_OPERAND (use, stmt, iter, SSA_OP_USE)
2131    {
2132      tree parm = unmodified_parm (stmt, use, NULL);
2133      /* For arguments we can build a condition.  */
2134      if (parm && ipa_get_param_decl_index (fbi->info, parm) >= 0)
2135	continue;
2136      if (TREE_CODE (use) != SSA_NAME)
2137	return p;
2138      /* If we know when operand is constant,
2139	 we still can say something useful.  */
2140      if (!true_predicate_p (&nonconstant_names[SSA_NAME_VERSION (use)]))
2141	continue;
2142      return p;
2143    }
2144
2145  if (is_load)
2146    op_non_const =
2147      add_condition (summary, base_index, size, &aggpos, CHANGED, NULL);
2148  else
2149    op_non_const = false_predicate ();
2150  FOR_EACH_SSA_TREE_OPERAND (use, stmt, iter, SSA_OP_USE)
2151    {
2152      HOST_WIDE_INT size;
2153      tree parm = unmodified_parm (stmt, use, &size);
2154      int index;
2155
2156      if (parm && (index = ipa_get_param_decl_index (fbi->info, parm)) >= 0)
2157	{
2158	  if (index != base_index)
2159	    p = add_condition (summary, index, size, NULL, CHANGED, NULL_TREE);
2160	  else
2161	    continue;
2162	}
2163      else
2164	p = nonconstant_names[SSA_NAME_VERSION (use)];
2165      op_non_const = or_predicates (summary->conds, &p, &op_non_const);
2166    }
2167  if ((gimple_code (stmt) == GIMPLE_ASSIGN || gimple_code (stmt) == GIMPLE_CALL)
2168      && gimple_op (stmt, 0)
2169      && TREE_CODE (gimple_op (stmt, 0)) == SSA_NAME)
2170    nonconstant_names[SSA_NAME_VERSION (gimple_op (stmt, 0))]
2171      = op_non_const;
2172  return op_non_const;
2173}
2174
2175struct record_modified_bb_info
2176{
2177  bitmap bb_set;
2178  gimple stmt;
2179};
2180
2181/* Callback of walk_aliased_vdefs.  Records basic blocks where the value may be
2182   set except for info->stmt.  */
2183
2184static bool
2185record_modified (ao_ref *ao ATTRIBUTE_UNUSED, tree vdef, void *data)
2186{
2187  struct record_modified_bb_info *info =
2188    (struct record_modified_bb_info *) data;
2189  if (SSA_NAME_DEF_STMT (vdef) == info->stmt)
2190    return false;
2191  bitmap_set_bit (info->bb_set,
2192		  SSA_NAME_IS_DEFAULT_DEF (vdef)
2193		  ? ENTRY_BLOCK_PTR_FOR_FN (cfun)->index
2194		  : gimple_bb (SSA_NAME_DEF_STMT (vdef))->index);
2195  return false;
2196}
2197
2198/* Return probability (based on REG_BR_PROB_BASE) that I-th parameter of STMT
2199   will change since last invocation of STMT.
2200
2201   Value 0 is reserved for compile time invariants.
2202   For common parameters it is REG_BR_PROB_BASE.  For loop invariants it
2203   ought to be REG_BR_PROB_BASE / estimated_iters.  */
2204
2205static int
2206param_change_prob (gimple stmt, int i)
2207{
2208  tree op = gimple_call_arg (stmt, i);
2209  basic_block bb = gimple_bb (stmt);
2210  tree base;
2211
2212  /* Global invariants neve change.  */
2213  if (is_gimple_min_invariant (op))
2214    return 0;
2215  /* We would have to do non-trivial analysis to really work out what
2216     is the probability of value to change (i.e. when init statement
2217     is in a sibling loop of the call).
2218
2219     We do an conservative estimate: when call is executed N times more often
2220     than the statement defining value, we take the frequency 1/N.  */
2221  if (TREE_CODE (op) == SSA_NAME)
2222    {
2223      int init_freq;
2224
2225      if (!bb->frequency)
2226	return REG_BR_PROB_BASE;
2227
2228      if (SSA_NAME_IS_DEFAULT_DEF (op))
2229	init_freq = ENTRY_BLOCK_PTR_FOR_FN (cfun)->frequency;
2230      else
2231	init_freq = gimple_bb (SSA_NAME_DEF_STMT (op))->frequency;
2232
2233      if (!init_freq)
2234	init_freq = 1;
2235      if (init_freq < bb->frequency)
2236	return MAX (GCOV_COMPUTE_SCALE (init_freq, bb->frequency), 1);
2237      else
2238	return REG_BR_PROB_BASE;
2239    }
2240
2241  base = get_base_address (op);
2242  if (base)
2243    {
2244      ao_ref refd;
2245      int max;
2246      struct record_modified_bb_info info;
2247      bitmap_iterator bi;
2248      unsigned index;
2249      tree init = ctor_for_folding (base);
2250
2251      if (init != error_mark_node)
2252	return 0;
2253      if (!bb->frequency)
2254	return REG_BR_PROB_BASE;
2255      ao_ref_init (&refd, op);
2256      info.stmt = stmt;
2257      info.bb_set = BITMAP_ALLOC (NULL);
2258      walk_aliased_vdefs (&refd, gimple_vuse (stmt), record_modified, &info,
2259			  NULL);
2260      if (bitmap_bit_p (info.bb_set, bb->index))
2261	{
2262	  BITMAP_FREE (info.bb_set);
2263	  return REG_BR_PROB_BASE;
2264	}
2265
2266      /* Assume that every memory is initialized at entry.
2267         TODO: Can we easilly determine if value is always defined
2268         and thus we may skip entry block?  */
2269      if (ENTRY_BLOCK_PTR_FOR_FN (cfun)->frequency)
2270	max = ENTRY_BLOCK_PTR_FOR_FN (cfun)->frequency;
2271      else
2272	max = 1;
2273
2274      EXECUTE_IF_SET_IN_BITMAP (info.bb_set, 0, index, bi)
2275	max = MIN (max, BASIC_BLOCK_FOR_FN (cfun, index)->frequency);
2276
2277      BITMAP_FREE (info.bb_set);
2278      if (max < bb->frequency)
2279	return MAX (GCOV_COMPUTE_SCALE (max, bb->frequency), 1);
2280      else
2281	return REG_BR_PROB_BASE;
2282    }
2283  return REG_BR_PROB_BASE;
2284}
2285
2286/* Find whether a basic block BB is the final block of a (half) diamond CFG
2287   sub-graph and if the predicate the condition depends on is known.  If so,
2288   return true and store the pointer the predicate in *P.  */
2289
2290static bool
2291phi_result_unknown_predicate (struct ipa_node_params *info,
2292			      inline_summary *summary, basic_block bb,
2293			      struct predicate *p,
2294			      vec<predicate_t> nonconstant_names)
2295{
2296  edge e;
2297  edge_iterator ei;
2298  basic_block first_bb = NULL;
2299  gimple stmt;
2300
2301  if (single_pred_p (bb))
2302    {
2303      *p = false_predicate ();
2304      return true;
2305    }
2306
2307  FOR_EACH_EDGE (e, ei, bb->preds)
2308    {
2309      if (single_succ_p (e->src))
2310	{
2311	  if (!single_pred_p (e->src))
2312	    return false;
2313	  if (!first_bb)
2314	    first_bb = single_pred (e->src);
2315	  else if (single_pred (e->src) != first_bb)
2316	    return false;
2317	}
2318      else
2319	{
2320	  if (!first_bb)
2321	    first_bb = e->src;
2322	  else if (e->src != first_bb)
2323	    return false;
2324	}
2325    }
2326
2327  if (!first_bb)
2328    return false;
2329
2330  stmt = last_stmt (first_bb);
2331  if (!stmt
2332      || gimple_code (stmt) != GIMPLE_COND
2333      || !is_gimple_ip_invariant (gimple_cond_rhs (stmt)))
2334    return false;
2335
2336  *p = will_be_nonconstant_expr_predicate (info, summary,
2337					   gimple_cond_lhs (stmt),
2338					   nonconstant_names);
2339  if (true_predicate_p (p))
2340    return false;
2341  else
2342    return true;
2343}
2344
2345/* Given a PHI statement in a function described by inline properties SUMMARY
2346   and *P being the predicate describing whether the selected PHI argument is
2347   known, store a predicate for the result of the PHI statement into
2348   NONCONSTANT_NAMES, if possible.  */
2349
2350static void
2351predicate_for_phi_result (struct inline_summary *summary, gphi *phi,
2352			  struct predicate *p,
2353			  vec<predicate_t> nonconstant_names)
2354{
2355  unsigned i;
2356
2357  for (i = 0; i < gimple_phi_num_args (phi); i++)
2358    {
2359      tree arg = gimple_phi_arg (phi, i)->def;
2360      if (!is_gimple_min_invariant (arg))
2361	{
2362	  gcc_assert (TREE_CODE (arg) == SSA_NAME);
2363	  *p = or_predicates (summary->conds, p,
2364			      &nonconstant_names[SSA_NAME_VERSION (arg)]);
2365	  if (true_predicate_p (p))
2366	    return;
2367	}
2368    }
2369
2370  if (dump_file && (dump_flags & TDF_DETAILS))
2371    {
2372      fprintf (dump_file, "\t\tphi predicate: ");
2373      dump_predicate (dump_file, summary->conds, p);
2374    }
2375  nonconstant_names[SSA_NAME_VERSION (gimple_phi_result (phi))] = *p;
2376}
2377
2378/* Return predicate specifying when array index in access OP becomes non-constant.  */
2379
2380static struct predicate
2381array_index_predicate (inline_summary *info,
2382		       vec< predicate_t> nonconstant_names, tree op)
2383{
2384  struct predicate p = false_predicate ();
2385  while (handled_component_p (op))
2386    {
2387      if (TREE_CODE (op) == ARRAY_REF || TREE_CODE (op) == ARRAY_RANGE_REF)
2388	{
2389	  if (TREE_CODE (TREE_OPERAND (op, 1)) == SSA_NAME)
2390	    p = or_predicates (info->conds, &p,
2391			       &nonconstant_names[SSA_NAME_VERSION
2392						  (TREE_OPERAND (op, 1))]);
2393	}
2394      op = TREE_OPERAND (op, 0);
2395    }
2396  return p;
2397}
2398
2399/* For a typical usage of __builtin_expect (a<b, 1), we
2400   may introduce an extra relation stmt:
2401   With the builtin, we have
2402     t1 = a <= b;
2403     t2 = (long int) t1;
2404     t3 = __builtin_expect (t2, 1);
2405     if (t3 != 0)
2406       goto ...
2407   Without the builtin, we have
2408     if (a<=b)
2409       goto...
2410   This affects the size/time estimation and may have
2411   an impact on the earlier inlining.
2412   Here find this pattern and fix it up later.  */
2413
2414static gimple
2415find_foldable_builtin_expect (basic_block bb)
2416{
2417  gimple_stmt_iterator bsi;
2418
2419  for (bsi = gsi_start_bb (bb); !gsi_end_p (bsi); gsi_next (&bsi))
2420    {
2421      gimple stmt = gsi_stmt (bsi);
2422      if (gimple_call_builtin_p (stmt, BUILT_IN_EXPECT)
2423	  || (is_gimple_call (stmt)
2424	      && gimple_call_internal_p (stmt)
2425	      && gimple_call_internal_fn (stmt) == IFN_BUILTIN_EXPECT))
2426        {
2427          tree var = gimple_call_lhs (stmt);
2428          tree arg = gimple_call_arg (stmt, 0);
2429          use_operand_p use_p;
2430          gimple use_stmt;
2431          bool match = false;
2432          bool done = false;
2433
2434          if (!var || !arg)
2435            continue;
2436          gcc_assert (TREE_CODE (var) == SSA_NAME);
2437
2438          while (TREE_CODE (arg) == SSA_NAME)
2439            {
2440              gimple stmt_tmp = SSA_NAME_DEF_STMT (arg);
2441              if (!is_gimple_assign (stmt_tmp))
2442                break;
2443              switch (gimple_assign_rhs_code (stmt_tmp))
2444                {
2445                  case LT_EXPR:
2446                  case LE_EXPR:
2447                  case GT_EXPR:
2448                  case GE_EXPR:
2449                  case EQ_EXPR:
2450                  case NE_EXPR:
2451                    match = true;
2452                    done = true;
2453                    break;
2454                  CASE_CONVERT:
2455                    break;
2456                  default:
2457                    done = true;
2458                    break;
2459                }
2460              if (done)
2461                break;
2462              arg = gimple_assign_rhs1 (stmt_tmp);
2463            }
2464
2465          if (match && single_imm_use (var, &use_p, &use_stmt)
2466              && gimple_code (use_stmt) == GIMPLE_COND)
2467            return use_stmt;
2468        }
2469    }
2470  return NULL;
2471}
2472
2473/* Return true when the basic blocks contains only clobbers followed by RESX.
2474   Such BBs are kept around to make removal of dead stores possible with
2475   presence of EH and will be optimized out by optimize_clobbers later in the
2476   game.
2477
2478   NEED_EH is used to recurse in case the clobber has non-EH predecestors
2479   that can be clobber only, too.. When it is false, the RESX is not necessary
2480   on the end of basic block.  */
2481
2482static bool
2483clobber_only_eh_bb_p (basic_block bb, bool need_eh = true)
2484{
2485  gimple_stmt_iterator gsi = gsi_last_bb (bb);
2486  edge_iterator ei;
2487  edge e;
2488
2489  if (need_eh)
2490    {
2491      if (gsi_end_p (gsi))
2492	return false;
2493      if (gimple_code (gsi_stmt (gsi)) != GIMPLE_RESX)
2494        return false;
2495      gsi_prev (&gsi);
2496    }
2497  else if (!single_succ_p (bb))
2498    return false;
2499
2500  for (; !gsi_end_p (gsi); gsi_prev (&gsi))
2501    {
2502      gimple stmt = gsi_stmt (gsi);
2503      if (is_gimple_debug (stmt))
2504	continue;
2505      if (gimple_clobber_p (stmt))
2506	continue;
2507      if (gimple_code (stmt) == GIMPLE_LABEL)
2508	break;
2509      return false;
2510    }
2511
2512  /* See if all predecestors are either throws or clobber only BBs.  */
2513  FOR_EACH_EDGE (e, ei, bb->preds)
2514    if (!(e->flags & EDGE_EH)
2515	&& !clobber_only_eh_bb_p (e->src, false))
2516      return false;
2517
2518  return true;
2519}
2520
2521/* Compute function body size parameters for NODE.
2522   When EARLY is true, we compute only simple summaries without
2523   non-trivial predicates to drive the early inliner.  */
2524
2525static void
2526estimate_function_body_sizes (struct cgraph_node *node, bool early)
2527{
2528  gcov_type time = 0;
2529  /* Estimate static overhead for function prologue/epilogue and alignment. */
2530  int size = 2;
2531  /* Benefits are scaled by probability of elimination that is in range
2532     <0,2>.  */
2533  basic_block bb;
2534  struct function *my_function = DECL_STRUCT_FUNCTION (node->decl);
2535  int freq;
2536  struct inline_summary *info = inline_summaries->get (node);
2537  struct predicate bb_predicate;
2538  struct ipa_func_body_info fbi;
2539  vec<predicate_t> nonconstant_names = vNULL;
2540  int nblocks, n;
2541  int *order;
2542  predicate array_index = true_predicate ();
2543  gimple fix_builtin_expect_stmt;
2544
2545  gcc_assert (my_function && my_function->cfg);
2546  gcc_assert (cfun == my_function);
2547
2548  memset(&fbi, 0, sizeof(fbi));
2549  info->conds = NULL;
2550  info->entry = NULL;
2551
2552  /* When optimizing and analyzing for IPA inliner, initialize loop optimizer
2553     so we can produce proper inline hints.
2554
2555     When optimizing and analyzing for early inliner, initialize node params
2556     so we can produce correct BB predicates.  */
2557
2558  if (opt_for_fn (node->decl, optimize))
2559    {
2560      calculate_dominance_info (CDI_DOMINATORS);
2561      if (!early)
2562        loop_optimizer_init (LOOPS_NORMAL | LOOPS_HAVE_RECORDED_EXITS);
2563      else
2564	{
2565	  ipa_check_create_node_params ();
2566	  ipa_initialize_node_params (node);
2567	}
2568
2569      if (ipa_node_params_sum)
2570	{
2571	  fbi.node = node;
2572	  fbi.info = IPA_NODE_REF (node);
2573	  fbi.bb_infos = vNULL;
2574	  fbi.bb_infos.safe_grow_cleared (last_basic_block_for_fn (cfun));
2575	  fbi.param_count = count_formal_params(node->decl);
2576	  nonconstant_names.safe_grow_cleared
2577	    (SSANAMES (my_function)->length ());
2578	}
2579    }
2580
2581  if (dump_file)
2582    fprintf (dump_file, "\nAnalyzing function body size: %s\n",
2583	     node->name ());
2584
2585  /* When we run into maximal number of entries, we assign everything to the
2586     constant truth case.  Be sure to have it in list. */
2587  bb_predicate = true_predicate ();
2588  account_size_time (info, 0, 0, &bb_predicate);
2589
2590  bb_predicate = not_inlined_predicate ();
2591  account_size_time (info, 2 * INLINE_SIZE_SCALE, 0, &bb_predicate);
2592
2593  if (fbi.info)
2594    compute_bb_predicates (&fbi, node, info);
2595  order = XNEWVEC (int, n_basic_blocks_for_fn (cfun));
2596  nblocks = pre_and_rev_post_order_compute (NULL, order, false);
2597  for (n = 0; n < nblocks; n++)
2598    {
2599      bb = BASIC_BLOCK_FOR_FN (cfun, order[n]);
2600      freq = compute_call_stmt_bb_frequency (node->decl, bb);
2601      if (clobber_only_eh_bb_p (bb))
2602	{
2603	  if (dump_file && (dump_flags & TDF_DETAILS))
2604	    fprintf (dump_file, "\n Ignoring BB %i;"
2605		     " it will be optimized away by cleanup_clobbers\n",
2606		     bb->index);
2607	  continue;
2608	}
2609
2610      /* TODO: Obviously predicates can be propagated down across CFG.  */
2611      if (fbi.info)
2612	{
2613	  if (bb->aux)
2614	    bb_predicate = *(struct predicate *) bb->aux;
2615	  else
2616	    bb_predicate = false_predicate ();
2617	}
2618      else
2619	bb_predicate = true_predicate ();
2620
2621      if (dump_file && (dump_flags & TDF_DETAILS))
2622	{
2623	  fprintf (dump_file, "\n BB %i predicate:", bb->index);
2624	  dump_predicate (dump_file, info->conds, &bb_predicate);
2625	}
2626
2627      if (fbi.info && nonconstant_names.exists ())
2628	{
2629	  struct predicate phi_predicate;
2630	  bool first_phi = true;
2631
2632	  for (gphi_iterator bsi = gsi_start_phis (bb); !gsi_end_p (bsi);
2633	       gsi_next (&bsi))
2634	    {
2635	      if (first_phi
2636		  && !phi_result_unknown_predicate (fbi.info, info, bb,
2637						    &phi_predicate,
2638						    nonconstant_names))
2639		break;
2640	      first_phi = false;
2641	      if (dump_file && (dump_flags & TDF_DETAILS))
2642		{
2643		  fprintf (dump_file, "  ");
2644		  print_gimple_stmt (dump_file, gsi_stmt (bsi), 0, 0);
2645		}
2646	      predicate_for_phi_result (info, bsi.phi (), &phi_predicate,
2647					nonconstant_names);
2648	    }
2649	}
2650
2651      fix_builtin_expect_stmt = find_foldable_builtin_expect (bb);
2652
2653      for (gimple_stmt_iterator bsi = gsi_start_bb (bb); !gsi_end_p (bsi);
2654	   gsi_next (&bsi))
2655	{
2656	  gimple stmt = gsi_stmt (bsi);
2657	  int this_size = estimate_num_insns (stmt, &eni_size_weights);
2658	  int this_time = estimate_num_insns (stmt, &eni_time_weights);
2659	  int prob;
2660	  struct predicate will_be_nonconstant;
2661
2662          /* This relation stmt should be folded after we remove
2663             buildin_expect call. Adjust the cost here.  */
2664	  if (stmt == fix_builtin_expect_stmt)
2665            {
2666              this_size--;
2667              this_time--;
2668            }
2669
2670	  if (dump_file && (dump_flags & TDF_DETAILS))
2671	    {
2672	      fprintf (dump_file, "  ");
2673	      print_gimple_stmt (dump_file, stmt, 0, 0);
2674	      fprintf (dump_file, "\t\tfreq:%3.2f size:%3i time:%3i\n",
2675		       ((double) freq) / CGRAPH_FREQ_BASE, this_size,
2676		       this_time);
2677	    }
2678
2679	  if (gimple_assign_load_p (stmt) && nonconstant_names.exists ())
2680	    {
2681	      struct predicate this_array_index;
2682	      this_array_index =
2683		array_index_predicate (info, nonconstant_names,
2684				       gimple_assign_rhs1 (stmt));
2685	      if (!false_predicate_p (&this_array_index))
2686		array_index =
2687		  and_predicates (info->conds, &array_index,
2688				  &this_array_index);
2689	    }
2690	  if (gimple_store_p (stmt) && nonconstant_names.exists ())
2691	    {
2692	      struct predicate this_array_index;
2693	      this_array_index =
2694		array_index_predicate (info, nonconstant_names,
2695				       gimple_get_lhs (stmt));
2696	      if (!false_predicate_p (&this_array_index))
2697		array_index =
2698		  and_predicates (info->conds, &array_index,
2699				  &this_array_index);
2700	    }
2701
2702
2703	  if (is_gimple_call (stmt)
2704	      && !gimple_call_internal_p (stmt))
2705	    {
2706	      struct cgraph_edge *edge = node->get_edge (stmt);
2707	      struct inline_edge_summary *es = inline_edge_summary (edge);
2708
2709	      /* Special case: results of BUILT_IN_CONSTANT_P will be always
2710	         resolved as constant.  We however don't want to optimize
2711	         out the cgraph edges.  */
2712	      if (nonconstant_names.exists ()
2713		  && gimple_call_builtin_p (stmt, BUILT_IN_CONSTANT_P)
2714		  && gimple_call_lhs (stmt)
2715		  && TREE_CODE (gimple_call_lhs (stmt)) == SSA_NAME)
2716		{
2717		  struct predicate false_p = false_predicate ();
2718		  nonconstant_names[SSA_NAME_VERSION (gimple_call_lhs (stmt))]
2719		    = false_p;
2720		}
2721	      if (ipa_node_params_sum)
2722		{
2723		  int count = gimple_call_num_args (stmt);
2724		  int i;
2725
2726		  if (count)
2727		    es->param.safe_grow_cleared (count);
2728		  for (i = 0; i < count; i++)
2729		    {
2730		      int prob = param_change_prob (stmt, i);
2731		      gcc_assert (prob >= 0 && prob <= REG_BR_PROB_BASE);
2732		      es->param[i].change_prob = prob;
2733		    }
2734		}
2735
2736	      es->call_stmt_size = this_size;
2737	      es->call_stmt_time = this_time;
2738	      es->loop_depth = bb_loop_depth (bb);
2739	      edge_set_predicate (edge, &bb_predicate);
2740	    }
2741
2742	  /* TODO: When conditional jump or swithc is known to be constant, but
2743	     we did not translate it into the predicates, we really can account
2744	     just maximum of the possible paths.  */
2745	  if (fbi.info)
2746	    will_be_nonconstant
2747	      = will_be_nonconstant_predicate (&fbi, info,
2748					       stmt, nonconstant_names);
2749	  if (this_time || this_size)
2750	    {
2751	      struct predicate p;
2752
2753	      this_time *= freq;
2754
2755	      prob = eliminated_by_inlining_prob (stmt);
2756	      if (prob == 1 && dump_file && (dump_flags & TDF_DETAILS))
2757		fprintf (dump_file,
2758			 "\t\t50%% will be eliminated by inlining\n");
2759	      if (prob == 2 && dump_file && (dump_flags & TDF_DETAILS))
2760		fprintf (dump_file, "\t\tWill be eliminated by inlining\n");
2761
2762	      if (fbi.info)
2763		p = and_predicates (info->conds, &bb_predicate,
2764				    &will_be_nonconstant);
2765	      else
2766		p = true_predicate ();
2767
2768	      if (!false_predicate_p (&p)
2769		  || (is_gimple_call (stmt)
2770		      && !false_predicate_p (&bb_predicate)))
2771		{
2772		  time += this_time;
2773		  size += this_size;
2774		  if (time > MAX_TIME * INLINE_TIME_SCALE)
2775		    time = MAX_TIME * INLINE_TIME_SCALE;
2776		}
2777
2778	      /* We account everything but the calls.  Calls have their own
2779	         size/time info attached to cgraph edges.  This is necessary
2780	         in order to make the cost disappear after inlining.  */
2781	      if (!is_gimple_call (stmt))
2782		{
2783		  if (prob)
2784		    {
2785		      struct predicate ip = not_inlined_predicate ();
2786		      ip = and_predicates (info->conds, &ip, &p);
2787		      account_size_time (info, this_size * prob,
2788					 this_time * prob, &ip);
2789		    }
2790		  if (prob != 2)
2791		    account_size_time (info, this_size * (2 - prob),
2792				       this_time * (2 - prob), &p);
2793		}
2794
2795	      gcc_assert (time >= 0);
2796	      gcc_assert (size >= 0);
2797	    }
2798	}
2799    }
2800  set_hint_predicate (&inline_summaries->get (node)->array_index, array_index);
2801  time = (time + CGRAPH_FREQ_BASE / 2) / CGRAPH_FREQ_BASE;
2802  if (time > MAX_TIME)
2803    time = MAX_TIME;
2804  free (order);
2805
2806  if (nonconstant_names.exists () && !early)
2807    {
2808      struct loop *loop;
2809      predicate loop_iterations = true_predicate ();
2810      predicate loop_stride = true_predicate ();
2811
2812      if (dump_file && (dump_flags & TDF_DETAILS))
2813	flow_loops_dump (dump_file, NULL, 0);
2814      scev_initialize ();
2815      FOR_EACH_LOOP (loop, 0)
2816	{
2817	  vec<edge> exits;
2818	  edge ex;
2819	  unsigned int j;
2820	  struct tree_niter_desc niter_desc;
2821	  bb_predicate = *(struct predicate *) loop->header->aux;
2822
2823	  exits = get_loop_exit_edges (loop);
2824	  FOR_EACH_VEC_ELT (exits, j, ex)
2825	    if (number_of_iterations_exit (loop, ex, &niter_desc, false)
2826		&& !is_gimple_min_invariant (niter_desc.niter))
2827	    {
2828	      predicate will_be_nonconstant
2829		= will_be_nonconstant_expr_predicate (fbi.info, info,
2830						      niter_desc.niter,
2831						      nonconstant_names);
2832	      if (!true_predicate_p (&will_be_nonconstant))
2833		will_be_nonconstant = and_predicates (info->conds,
2834						      &bb_predicate,
2835						      &will_be_nonconstant);
2836	      if (!true_predicate_p (&will_be_nonconstant)
2837		  && !false_predicate_p (&will_be_nonconstant))
2838		/* This is slightly inprecise.  We may want to represent each
2839		   loop with independent predicate.  */
2840		loop_iterations =
2841		  and_predicates (info->conds, &loop_iterations,
2842				  &will_be_nonconstant);
2843	    }
2844	  exits.release ();
2845
2846	  for (gphi_iterator gsi = gsi_start_phis (loop->header);
2847	       !gsi_end_p (gsi); gsi_next (&gsi))
2848	    {
2849	      gphi *phi = gsi.phi ();
2850	      tree use = gimple_phi_result (phi);
2851	      affine_iv iv;
2852	      predicate will_be_nonconstant;
2853	      if (virtual_operand_p (use)
2854		  || !simple_iv (loop, loop, use, &iv, true)
2855		  || is_gimple_min_invariant (iv.step))
2856		continue;
2857	      will_be_nonconstant
2858		= will_be_nonconstant_expr_predicate (fbi.info, info,
2859						      iv.step,
2860						      nonconstant_names);
2861	      if (!true_predicate_p (&will_be_nonconstant))
2862		will_be_nonconstant = and_predicates (info->conds,
2863						      &bb_predicate,
2864						      &will_be_nonconstant);
2865	      if (!true_predicate_p (&will_be_nonconstant)
2866		  && !false_predicate_p (&will_be_nonconstant))
2867		/* This is slightly inprecise.  We may want to represent
2868		   each loop with independent predicate.  */
2869		loop_stride = and_predicates (info->conds, &loop_stride,
2870					      &will_be_nonconstant);
2871	    }
2872	}
2873      set_hint_predicate (&inline_summaries->get (node)->loop_iterations,
2874			  loop_iterations);
2875      set_hint_predicate (&inline_summaries->get (node)->loop_stride, loop_stride);
2876      scev_finalize ();
2877    }
2878  FOR_ALL_BB_FN (bb, my_function)
2879    {
2880      edge e;
2881      edge_iterator ei;
2882
2883      if (bb->aux)
2884	pool_free (edge_predicate_pool, bb->aux);
2885      bb->aux = NULL;
2886      FOR_EACH_EDGE (e, ei, bb->succs)
2887	{
2888	  if (e->aux)
2889	    pool_free (edge_predicate_pool, e->aux);
2890	  e->aux = NULL;
2891	}
2892    }
2893  inline_summaries->get (node)->self_time = time;
2894  inline_summaries->get (node)->self_size = size;
2895  nonconstant_names.release ();
2896  if (opt_for_fn (node->decl, optimize))
2897    {
2898      if (!early)
2899        loop_optimizer_finalize ();
2900      else if (!ipa_edge_args_vector)
2901	ipa_free_all_node_params ();
2902      free_dominance_info (CDI_DOMINATORS);
2903    }
2904  if (dump_file)
2905    {
2906      fprintf (dump_file, "\n");
2907      dump_inline_summary (dump_file, node);
2908    }
2909}
2910
2911
2912/* Compute parameters of functions used by inliner.
2913   EARLY is true when we compute parameters for the early inliner  */
2914
2915void
2916compute_inline_parameters (struct cgraph_node *node, bool early)
2917{
2918  HOST_WIDE_INT self_stack_size;
2919  struct cgraph_edge *e;
2920  struct inline_summary *info;
2921
2922  gcc_assert (!node->global.inlined_to);
2923
2924  inline_summary_alloc ();
2925
2926  info = inline_summaries->get (node);
2927  reset_inline_summary (node, info);
2928
2929  /* FIXME: Thunks are inlinable, but tree-inline don't know how to do that.
2930     Once this happen, we will need to more curefully predict call
2931     statement size.  */
2932  if (node->thunk.thunk_p)
2933    {
2934      struct inline_edge_summary *es = inline_edge_summary (node->callees);
2935      struct predicate t = true_predicate ();
2936
2937      info->inlinable = 0;
2938      node->callees->call_stmt_cannot_inline_p = true;
2939      node->local.can_change_signature = false;
2940      es->call_stmt_time = 1;
2941      es->call_stmt_size = 1;
2942      account_size_time (info, 0, 0, &t);
2943      return;
2944    }
2945
2946  /* Even is_gimple_min_invariant rely on current_function_decl.  */
2947  push_cfun (DECL_STRUCT_FUNCTION (node->decl));
2948
2949  /* Estimate the stack size for the function if we're optimizing.  */
2950  self_stack_size = optimize ? estimated_stack_frame_size (node) : 0;
2951  info->estimated_self_stack_size = self_stack_size;
2952  info->estimated_stack_size = self_stack_size;
2953  info->stack_frame_offset = 0;
2954
2955  /* Can this function be inlined at all?  */
2956  if (!opt_for_fn (node->decl, optimize)
2957      && !lookup_attribute ("always_inline",
2958			    DECL_ATTRIBUTES (node->decl)))
2959    info->inlinable = false;
2960  else
2961    info->inlinable = tree_inlinable_function_p (node->decl);
2962
2963  info->contains_cilk_spawn = fn_contains_cilk_spawn_p (cfun);
2964
2965  /* Type attributes can use parameter indices to describe them.  */
2966  if (TYPE_ATTRIBUTES (TREE_TYPE (node->decl)))
2967    node->local.can_change_signature = false;
2968  else
2969    {
2970      /* Otherwise, inlinable functions always can change signature.  */
2971      if (info->inlinable)
2972	node->local.can_change_signature = true;
2973      else
2974	{
2975	  /* Functions calling builtin_apply can not change signature.  */
2976	  for (e = node->callees; e; e = e->next_callee)
2977	    {
2978	      tree cdecl = e->callee->decl;
2979	      if (DECL_BUILT_IN (cdecl)
2980		  && DECL_BUILT_IN_CLASS (cdecl) == BUILT_IN_NORMAL
2981		  && (DECL_FUNCTION_CODE (cdecl) == BUILT_IN_APPLY_ARGS
2982		      || DECL_FUNCTION_CODE (cdecl) == BUILT_IN_VA_START))
2983		break;
2984	    }
2985	  node->local.can_change_signature = !e;
2986	}
2987    }
2988  estimate_function_body_sizes (node, early);
2989
2990  for (e = node->callees; e; e = e->next_callee)
2991    if (e->callee->comdat_local_p ())
2992      break;
2993  node->calls_comdat_local = (e != NULL);
2994
2995  /* Inlining characteristics are maintained by the cgraph_mark_inline.  */
2996  info->time = info->self_time;
2997  info->size = info->self_size;
2998  info->stack_frame_offset = 0;
2999  info->estimated_stack_size = info->estimated_self_stack_size;
3000#ifdef ENABLE_CHECKING
3001  inline_update_overall_summary (node);
3002  gcc_assert (info->time == info->self_time && info->size == info->self_size);
3003#endif
3004
3005  pop_cfun ();
3006}
3007
3008
3009/* Compute parameters of functions used by inliner using
3010   current_function_decl.  */
3011
3012static unsigned int
3013compute_inline_parameters_for_current (void)
3014{
3015  compute_inline_parameters (cgraph_node::get (current_function_decl), true);
3016  return 0;
3017}
3018
3019namespace {
3020
3021const pass_data pass_data_inline_parameters =
3022{
3023  GIMPLE_PASS, /* type */
3024  "inline_param", /* name */
3025  OPTGROUP_INLINE, /* optinfo_flags */
3026  TV_INLINE_PARAMETERS, /* tv_id */
3027  0, /* properties_required */
3028  0, /* properties_provided */
3029  0, /* properties_destroyed */
3030  0, /* todo_flags_start */
3031  0, /* todo_flags_finish */
3032};
3033
3034class pass_inline_parameters : public gimple_opt_pass
3035{
3036public:
3037  pass_inline_parameters (gcc::context *ctxt)
3038    : gimple_opt_pass (pass_data_inline_parameters, ctxt)
3039  {}
3040
3041  /* opt_pass methods: */
3042  opt_pass * clone () { return new pass_inline_parameters (m_ctxt); }
3043  virtual unsigned int execute (function *)
3044    {
3045      return compute_inline_parameters_for_current ();
3046    }
3047
3048}; // class pass_inline_parameters
3049
3050} // anon namespace
3051
3052gimple_opt_pass *
3053make_pass_inline_parameters (gcc::context *ctxt)
3054{
3055  return new pass_inline_parameters (ctxt);
3056}
3057
3058
3059/* Estimate benefit devirtualizing indirect edge IE, provided KNOWN_VALS,
3060   KNOWN_CONTEXTS and KNOWN_AGGS.  */
3061
3062static bool
3063estimate_edge_devirt_benefit (struct cgraph_edge *ie,
3064			      int *size, int *time,
3065			      vec<tree> known_vals,
3066			      vec<ipa_polymorphic_call_context> known_contexts,
3067			      vec<ipa_agg_jump_function_p> known_aggs)
3068{
3069  tree target;
3070  struct cgraph_node *callee;
3071  struct inline_summary *isummary;
3072  enum availability avail;
3073  bool speculative;
3074
3075  if (!known_vals.exists () && !known_contexts.exists ())
3076    return false;
3077  if (!opt_for_fn (ie->caller->decl, flag_indirect_inlining))
3078    return false;
3079
3080  target = ipa_get_indirect_edge_target (ie, known_vals, known_contexts,
3081					 known_aggs, &speculative);
3082  if (!target || speculative)
3083    return false;
3084
3085  /* Account for difference in cost between indirect and direct calls.  */
3086  *size -= (eni_size_weights.indirect_call_cost - eni_size_weights.call_cost);
3087  *time -= (eni_time_weights.indirect_call_cost - eni_time_weights.call_cost);
3088  gcc_checking_assert (*time >= 0);
3089  gcc_checking_assert (*size >= 0);
3090
3091  callee = cgraph_node::get (target);
3092  if (!callee || !callee->definition)
3093    return false;
3094  callee = callee->function_symbol (&avail);
3095  if (avail < AVAIL_AVAILABLE)
3096    return false;
3097  isummary = inline_summaries->get (callee);
3098  return isummary->inlinable;
3099}
3100
3101/* Increase SIZE, MIN_SIZE (if non-NULL) and TIME for size and time needed to
3102   handle edge E with probability PROB.
3103   Set HINTS if edge may be devirtualized.
3104   KNOWN_VALS, KNOWN_AGGS and KNOWN_CONTEXTS describe context of the call
3105   site.  */
3106
3107static inline void
3108estimate_edge_size_and_time (struct cgraph_edge *e, int *size, int *min_size,
3109			     int *time,
3110			     int prob,
3111			     vec<tree> known_vals,
3112			     vec<ipa_polymorphic_call_context> known_contexts,
3113			     vec<ipa_agg_jump_function_p> known_aggs,
3114			     inline_hints *hints)
3115{
3116  struct inline_edge_summary *es = inline_edge_summary (e);
3117  int call_size = es->call_stmt_size;
3118  int call_time = es->call_stmt_time;
3119  int cur_size;
3120  if (!e->callee
3121      && estimate_edge_devirt_benefit (e, &call_size, &call_time,
3122				       known_vals, known_contexts, known_aggs)
3123      && hints && e->maybe_hot_p ())
3124    *hints |= INLINE_HINT_indirect_call;
3125  cur_size = call_size * INLINE_SIZE_SCALE;
3126  *size += cur_size;
3127  if (min_size)
3128    *min_size += cur_size;
3129  *time += apply_probability ((gcov_type) call_time, prob)
3130    * e->frequency * (INLINE_TIME_SCALE / CGRAPH_FREQ_BASE);
3131  if (*time > MAX_TIME * INLINE_TIME_SCALE)
3132    *time = MAX_TIME * INLINE_TIME_SCALE;
3133}
3134
3135
3136
3137/* Increase SIZE, MIN_SIZE and TIME for size and time needed to handle all
3138   calls in NODE.  POSSIBLE_TRUTHS, KNOWN_VALS, KNOWN_AGGS and KNOWN_CONTEXTS
3139   describe context of the call site.  */
3140
3141static void
3142estimate_calls_size_and_time (struct cgraph_node *node, int *size,
3143			      int *min_size, int *time,
3144			      inline_hints *hints,
3145			      clause_t possible_truths,
3146			      vec<tree> known_vals,
3147			      vec<ipa_polymorphic_call_context> known_contexts,
3148			      vec<ipa_agg_jump_function_p> known_aggs)
3149{
3150  struct cgraph_edge *e;
3151  for (e = node->callees; e; e = e->next_callee)
3152    {
3153      if (inline_edge_summary_vec.length () <= (unsigned) e->uid)
3154	continue;
3155
3156      struct inline_edge_summary *es = inline_edge_summary (e);
3157
3158      /* Do not care about zero sized builtins.  */
3159      if (e->inline_failed && !es->call_stmt_size)
3160	{
3161	  gcc_checking_assert (!es->call_stmt_time);
3162	  continue;
3163	}
3164      if (!es->predicate
3165	  || evaluate_predicate (es->predicate, possible_truths))
3166	{
3167	  if (e->inline_failed)
3168	    {
3169	      /* Predicates of calls shall not use NOT_CHANGED codes,
3170	         sowe do not need to compute probabilities.  */
3171	      estimate_edge_size_and_time (e, size,
3172					   es->predicate ? NULL : min_size,
3173					   time, REG_BR_PROB_BASE,
3174					   known_vals, known_contexts,
3175					   known_aggs, hints);
3176	    }
3177	  else
3178	    estimate_calls_size_and_time (e->callee, size, min_size, time,
3179					  hints,
3180					  possible_truths,
3181					  known_vals, known_contexts,
3182					  known_aggs);
3183	}
3184    }
3185  for (e = node->indirect_calls; e; e = e->next_callee)
3186    {
3187      if (inline_edge_summary_vec.length () <= (unsigned) e->uid)
3188	continue;
3189
3190      struct inline_edge_summary *es = inline_edge_summary (e);
3191      if (!es->predicate
3192	  || evaluate_predicate (es->predicate, possible_truths))
3193	estimate_edge_size_and_time (e, size,
3194				     es->predicate ? NULL : min_size,
3195				     time, REG_BR_PROB_BASE,
3196				     known_vals, known_contexts, known_aggs,
3197				     hints);
3198    }
3199}
3200
3201
3202/* Estimate size and time needed to execute NODE assuming
3203   POSSIBLE_TRUTHS clause, and KNOWN_VALS, KNOWN_AGGS and KNOWN_CONTEXTS
3204   information about NODE's arguments.  If non-NULL use also probability
3205   information present in INLINE_PARAM_SUMMARY vector.
3206   Additionally detemine hints determined by the context.  Finally compute
3207   minimal size needed for the call that is independent on the call context and
3208   can be used for fast estimates.  Return the values in RET_SIZE,
3209   RET_MIN_SIZE, RET_TIME and RET_HINTS.  */
3210
3211static void
3212estimate_node_size_and_time (struct cgraph_node *node,
3213			     clause_t possible_truths,
3214			     vec<tree> known_vals,
3215			     vec<ipa_polymorphic_call_context> known_contexts,
3216			     vec<ipa_agg_jump_function_p> known_aggs,
3217			     int *ret_size, int *ret_min_size, int *ret_time,
3218			     inline_hints *ret_hints,
3219			     vec<inline_param_summary>
3220			     inline_param_summary)
3221{
3222  struct inline_summary *info = inline_summaries->get (node);
3223  size_time_entry *e;
3224  int size = 0;
3225  int time = 0;
3226  int min_size = 0;
3227  inline_hints hints = 0;
3228  int i;
3229
3230  if (dump_file && (dump_flags & TDF_DETAILS))
3231    {
3232      bool found = false;
3233      fprintf (dump_file, "   Estimating body: %s/%i\n"
3234	       "   Known to be false: ", node->name (),
3235	       node->order);
3236
3237      for (i = predicate_not_inlined_condition;
3238	   i < (predicate_first_dynamic_condition
3239		+ (int) vec_safe_length (info->conds)); i++)
3240	if (!(possible_truths & (1 << i)))
3241	  {
3242	    if (found)
3243	      fprintf (dump_file, ", ");
3244	    found = true;
3245	    dump_condition (dump_file, info->conds, i);
3246	  }
3247    }
3248
3249  for (i = 0; vec_safe_iterate (info->entry, i, &e); i++)
3250    if (evaluate_predicate (&e->predicate, possible_truths))
3251      {
3252	size += e->size;
3253	gcc_checking_assert (e->time >= 0);
3254	gcc_checking_assert (time >= 0);
3255	if (!inline_param_summary.exists ())
3256	  time += e->time;
3257	else
3258	  {
3259	    int prob = predicate_probability (info->conds,
3260					      &e->predicate,
3261					      possible_truths,
3262					      inline_param_summary);
3263	    gcc_checking_assert (prob >= 0);
3264	    gcc_checking_assert (prob <= REG_BR_PROB_BASE);
3265	    time += apply_probability ((gcov_type) e->time, prob);
3266	  }
3267	if (time > MAX_TIME * INLINE_TIME_SCALE)
3268	  time = MAX_TIME * INLINE_TIME_SCALE;
3269	gcc_checking_assert (time >= 0);
3270
3271      }
3272  gcc_checking_assert (true_predicate_p (&(*info->entry)[0].predicate));
3273  min_size = (*info->entry)[0].size;
3274  gcc_checking_assert (size >= 0);
3275  gcc_checking_assert (time >= 0);
3276
3277  if (info->loop_iterations
3278      && !evaluate_predicate (info->loop_iterations, possible_truths))
3279    hints |= INLINE_HINT_loop_iterations;
3280  if (info->loop_stride
3281      && !evaluate_predicate (info->loop_stride, possible_truths))
3282    hints |= INLINE_HINT_loop_stride;
3283  if (info->array_index
3284      && !evaluate_predicate (info->array_index, possible_truths))
3285    hints |= INLINE_HINT_array_index;
3286  if (info->scc_no)
3287    hints |= INLINE_HINT_in_scc;
3288  if (DECL_DECLARED_INLINE_P (node->decl))
3289    hints |= INLINE_HINT_declared_inline;
3290
3291  estimate_calls_size_and_time (node, &size, &min_size, &time, &hints, possible_truths,
3292				known_vals, known_contexts, known_aggs);
3293  gcc_checking_assert (size >= 0);
3294  gcc_checking_assert (time >= 0);
3295  time = RDIV (time, INLINE_TIME_SCALE);
3296  size = RDIV (size, INLINE_SIZE_SCALE);
3297  min_size = RDIV (min_size, INLINE_SIZE_SCALE);
3298
3299  if (dump_file && (dump_flags & TDF_DETAILS))
3300    fprintf (dump_file, "\n   size:%i time:%i\n", (int) size, (int) time);
3301  if (ret_time)
3302    *ret_time = time;
3303  if (ret_size)
3304    *ret_size = size;
3305  if (ret_min_size)
3306    *ret_min_size = min_size;
3307  if (ret_hints)
3308    *ret_hints = hints;
3309  return;
3310}
3311
3312
3313/* Estimate size and time needed to execute callee of EDGE assuming that
3314   parameters known to be constant at caller of EDGE are propagated.
3315   KNOWN_VALS and KNOWN_CONTEXTS are vectors of assumed known constant values
3316   and types for parameters.  */
3317
3318void
3319estimate_ipcp_clone_size_and_time (struct cgraph_node *node,
3320				   vec<tree> known_vals,
3321				   vec<ipa_polymorphic_call_context>
3322				   known_contexts,
3323				   vec<ipa_agg_jump_function_p> known_aggs,
3324				   int *ret_size, int *ret_time,
3325				   inline_hints *hints)
3326{
3327  clause_t clause;
3328
3329  clause = evaluate_conditions_for_known_args (node, false, known_vals,
3330					       known_aggs);
3331  estimate_node_size_and_time (node, clause, known_vals, known_contexts,
3332			       known_aggs, ret_size, NULL, ret_time, hints, vNULL);
3333}
3334
3335/* Translate all conditions from callee representation into caller
3336   representation and symbolically evaluate predicate P into new predicate.
3337
3338   INFO is inline_summary of function we are adding predicate into, CALLEE_INFO
3339   is summary of function predicate P is from. OPERAND_MAP is array giving
3340   callee formal IDs the caller formal IDs. POSSSIBLE_TRUTHS is clausule of all
3341   callee conditions that may be true in caller context.  TOPLEV_PREDICATE is
3342   predicate under which callee is executed.  OFFSET_MAP is an array of of
3343   offsets that need to be added to conditions, negative offset means that
3344   conditions relying on values passed by reference have to be discarded
3345   because they might not be preserved (and should be considered offset zero
3346   for other purposes).  */
3347
3348static struct predicate
3349remap_predicate (struct inline_summary *info,
3350		 struct inline_summary *callee_info,
3351		 struct predicate *p,
3352		 vec<int> operand_map,
3353		 vec<int> offset_map,
3354		 clause_t possible_truths, struct predicate *toplev_predicate)
3355{
3356  int i;
3357  struct predicate out = true_predicate ();
3358
3359  /* True predicate is easy.  */
3360  if (true_predicate_p (p))
3361    return *toplev_predicate;
3362  for (i = 0; p->clause[i]; i++)
3363    {
3364      clause_t clause = p->clause[i];
3365      int cond;
3366      struct predicate clause_predicate = false_predicate ();
3367
3368      gcc_assert (i < MAX_CLAUSES);
3369
3370      for (cond = 0; cond < NUM_CONDITIONS; cond++)
3371	/* Do we have condition we can't disprove?   */
3372	if (clause & possible_truths & (1 << cond))
3373	  {
3374	    struct predicate cond_predicate;
3375	    /* Work out if the condition can translate to predicate in the
3376	       inlined function.  */
3377	    if (cond >= predicate_first_dynamic_condition)
3378	      {
3379		struct condition *c;
3380
3381		c = &(*callee_info->conds)[cond
3382					   -
3383					   predicate_first_dynamic_condition];
3384		/* See if we can remap condition operand to caller's operand.
3385		   Otherwise give up.  */
3386		if (!operand_map.exists ()
3387		    || (int) operand_map.length () <= c->operand_num
3388		    || operand_map[c->operand_num] == -1
3389		    /* TODO: For non-aggregate conditions, adding an offset is
3390		       basically an arithmetic jump function processing which
3391		       we should support in future.  */
3392		    || ((!c->agg_contents || !c->by_ref)
3393			&& offset_map[c->operand_num] > 0)
3394		    || (c->agg_contents && c->by_ref
3395			&& offset_map[c->operand_num] < 0))
3396		  cond_predicate = true_predicate ();
3397		else
3398		  {
3399		    struct agg_position_info ap;
3400		    HOST_WIDE_INT offset_delta = offset_map[c->operand_num];
3401		    if (offset_delta < 0)
3402		      {
3403			gcc_checking_assert (!c->agg_contents || !c->by_ref);
3404			offset_delta = 0;
3405		      }
3406		    gcc_assert (!c->agg_contents
3407				|| c->by_ref || offset_delta == 0);
3408		    ap.offset = c->offset + offset_delta;
3409		    ap.agg_contents = c->agg_contents;
3410		    ap.by_ref = c->by_ref;
3411		    cond_predicate = add_condition (info,
3412						    operand_map[c->operand_num],
3413						    c->size, &ap, c->code,
3414						    c->val);
3415		  }
3416	      }
3417	    /* Fixed conditions remains same, construct single
3418	       condition predicate.  */
3419	    else
3420	      {
3421		cond_predicate.clause[0] = 1 << cond;
3422		cond_predicate.clause[1] = 0;
3423	      }
3424	    clause_predicate = or_predicates (info->conds, &clause_predicate,
3425					      &cond_predicate);
3426	  }
3427      out = and_predicates (info->conds, &out, &clause_predicate);
3428    }
3429  return and_predicates (info->conds, &out, toplev_predicate);
3430}
3431
3432
3433/* Update summary information of inline clones after inlining.
3434   Compute peak stack usage.  */
3435
3436static void
3437inline_update_callee_summaries (struct cgraph_node *node, int depth)
3438{
3439  struct cgraph_edge *e;
3440  struct inline_summary *callee_info = inline_summaries->get (node);
3441  struct inline_summary *caller_info = inline_summaries->get (node->callers->caller);
3442  HOST_WIDE_INT peak;
3443
3444  callee_info->stack_frame_offset
3445    = caller_info->stack_frame_offset
3446    + caller_info->estimated_self_stack_size;
3447  peak = callee_info->stack_frame_offset
3448    + callee_info->estimated_self_stack_size;
3449  if (inline_summaries->get (node->global.inlined_to)->estimated_stack_size < peak)
3450      inline_summaries->get (node->global.inlined_to)->estimated_stack_size = peak;
3451  ipa_propagate_frequency (node);
3452  for (e = node->callees; e; e = e->next_callee)
3453    {
3454      if (!e->inline_failed)
3455	inline_update_callee_summaries (e->callee, depth);
3456      inline_edge_summary (e)->loop_depth += depth;
3457    }
3458  for (e = node->indirect_calls; e; e = e->next_callee)
3459    inline_edge_summary (e)->loop_depth += depth;
3460}
3461
3462/* Update change_prob of EDGE after INLINED_EDGE has been inlined.
3463   When functoin A is inlined in B and A calls C with parameter that
3464   changes with probability PROB1 and C is known to be passthroug
3465   of argument if B that change with probability PROB2, the probability
3466   of change is now PROB1*PROB2.  */
3467
3468static void
3469remap_edge_change_prob (struct cgraph_edge *inlined_edge,
3470			struct cgraph_edge *edge)
3471{
3472  if (ipa_node_params_sum)
3473    {
3474      int i;
3475      struct ipa_edge_args *args = IPA_EDGE_REF (edge);
3476      struct inline_edge_summary *es = inline_edge_summary (edge);
3477      struct inline_edge_summary *inlined_es
3478	= inline_edge_summary (inlined_edge);
3479
3480      for (i = 0; i < ipa_get_cs_argument_count (args); i++)
3481	{
3482	  struct ipa_jump_func *jfunc = ipa_get_ith_jump_func (args, i);
3483	  if (jfunc->type == IPA_JF_PASS_THROUGH
3484	      && (ipa_get_jf_pass_through_formal_id (jfunc)
3485		  < (int) inlined_es->param.length ()))
3486	    {
3487	      int jf_formal_id = ipa_get_jf_pass_through_formal_id (jfunc);
3488	      int prob1 = es->param[i].change_prob;
3489	      int prob2 = inlined_es->param[jf_formal_id].change_prob;
3490	      int prob = combine_probabilities (prob1, prob2);
3491
3492	      if (prob1 && prob2 && !prob)
3493		prob = 1;
3494
3495	      es->param[i].change_prob = prob;
3496	    }
3497	}
3498    }
3499}
3500
3501/* Update edge summaries of NODE after INLINED_EDGE has been inlined.
3502
3503   Remap predicates of callees of NODE.  Rest of arguments match
3504   remap_predicate.
3505
3506   Also update change probabilities.  */
3507
3508static void
3509remap_edge_summaries (struct cgraph_edge *inlined_edge,
3510		      struct cgraph_node *node,
3511		      struct inline_summary *info,
3512		      struct inline_summary *callee_info,
3513		      vec<int> operand_map,
3514		      vec<int> offset_map,
3515		      clause_t possible_truths,
3516		      struct predicate *toplev_predicate)
3517{
3518  struct cgraph_edge *e, *next;
3519  for (e = node->callees; e; e = next)
3520    {
3521      struct inline_edge_summary *es = inline_edge_summary (e);
3522      struct predicate p;
3523      next = e->next_callee;
3524
3525      if (e->inline_failed)
3526	{
3527	  remap_edge_change_prob (inlined_edge, e);
3528
3529	  if (es->predicate)
3530	    {
3531	      p = remap_predicate (info, callee_info,
3532				   es->predicate, operand_map, offset_map,
3533				   possible_truths, toplev_predicate);
3534	      edge_set_predicate (e, &p);
3535	    }
3536	  else
3537	    edge_set_predicate (e, toplev_predicate);
3538	}
3539      else
3540	remap_edge_summaries (inlined_edge, e->callee, info, callee_info,
3541			      operand_map, offset_map, possible_truths,
3542			      toplev_predicate);
3543    }
3544  for (e = node->indirect_calls; e; e = next)
3545    {
3546      struct inline_edge_summary *es = inline_edge_summary (e);
3547      struct predicate p;
3548      next = e->next_callee;
3549
3550      remap_edge_change_prob (inlined_edge, e);
3551      if (es->predicate)
3552	{
3553	  p = remap_predicate (info, callee_info,
3554			       es->predicate, operand_map, offset_map,
3555			       possible_truths, toplev_predicate);
3556	  edge_set_predicate (e, &p);
3557	}
3558      else
3559	edge_set_predicate (e, toplev_predicate);
3560    }
3561}
3562
3563/* Same as remap_predicate, but set result into hint *HINT.  */
3564
3565static void
3566remap_hint_predicate (struct inline_summary *info,
3567		      struct inline_summary *callee_info,
3568		      struct predicate **hint,
3569		      vec<int> operand_map,
3570		      vec<int> offset_map,
3571		      clause_t possible_truths,
3572		      struct predicate *toplev_predicate)
3573{
3574  predicate p;
3575
3576  if (!*hint)
3577    return;
3578  p = remap_predicate (info, callee_info,
3579		       *hint,
3580		       operand_map, offset_map,
3581		       possible_truths, toplev_predicate);
3582  if (!false_predicate_p (&p) && !true_predicate_p (&p))
3583    {
3584      if (!*hint)
3585	set_hint_predicate (hint, p);
3586      else
3587	**hint = and_predicates (info->conds, *hint, &p);
3588    }
3589}
3590
3591/* We inlined EDGE.  Update summary of the function we inlined into.  */
3592
3593void
3594inline_merge_summary (struct cgraph_edge *edge)
3595{
3596  struct inline_summary *callee_info = inline_summaries->get (edge->callee);
3597  struct cgraph_node *to = (edge->caller->global.inlined_to
3598			    ? edge->caller->global.inlined_to : edge->caller);
3599  struct inline_summary *info = inline_summaries->get (to);
3600  clause_t clause = 0;		/* not_inline is known to be false.  */
3601  size_time_entry *e;
3602  vec<int> operand_map = vNULL;
3603  vec<int> offset_map = vNULL;
3604  int i;
3605  struct predicate toplev_predicate;
3606  struct predicate true_p = true_predicate ();
3607  struct inline_edge_summary *es = inline_edge_summary (edge);
3608
3609  if (es->predicate)
3610    toplev_predicate = *es->predicate;
3611  else
3612    toplev_predicate = true_predicate ();
3613
3614  if (callee_info->conds)
3615    evaluate_properties_for_edge (edge, true, &clause, NULL, NULL, NULL);
3616  if (ipa_node_params_sum && callee_info->conds)
3617    {
3618      struct ipa_edge_args *args = IPA_EDGE_REF (edge);
3619      int count = ipa_get_cs_argument_count (args);
3620      int i;
3621
3622      if (count)
3623	{
3624	  operand_map.safe_grow_cleared (count);
3625	  offset_map.safe_grow_cleared (count);
3626	}
3627      for (i = 0; i < count; i++)
3628	{
3629	  struct ipa_jump_func *jfunc = ipa_get_ith_jump_func (args, i);
3630	  int map = -1;
3631
3632	  /* TODO: handle non-NOPs when merging.  */
3633	  if (jfunc->type == IPA_JF_PASS_THROUGH)
3634	    {
3635	      if (ipa_get_jf_pass_through_operation (jfunc) == NOP_EXPR)
3636		map = ipa_get_jf_pass_through_formal_id (jfunc);
3637	      if (!ipa_get_jf_pass_through_agg_preserved (jfunc))
3638		offset_map[i] = -1;
3639	    }
3640	  else if (jfunc->type == IPA_JF_ANCESTOR)
3641	    {
3642	      HOST_WIDE_INT offset = ipa_get_jf_ancestor_offset (jfunc);
3643	      if (offset >= 0 && offset < INT_MAX)
3644		{
3645		  map = ipa_get_jf_ancestor_formal_id (jfunc);
3646		  if (!ipa_get_jf_ancestor_agg_preserved (jfunc))
3647		    offset = -1;
3648		  offset_map[i] = offset;
3649		}
3650	    }
3651	  operand_map[i] = map;
3652	  gcc_assert (map < ipa_get_param_count (IPA_NODE_REF (to)));
3653	}
3654    }
3655  for (i = 0; vec_safe_iterate (callee_info->entry, i, &e); i++)
3656    {
3657      struct predicate p = remap_predicate (info, callee_info,
3658					    &e->predicate, operand_map,
3659					    offset_map, clause,
3660					    &toplev_predicate);
3661      if (!false_predicate_p (&p))
3662	{
3663	  gcov_type add_time = ((gcov_type) e->time * edge->frequency
3664				+ CGRAPH_FREQ_BASE / 2) / CGRAPH_FREQ_BASE;
3665	  int prob = predicate_probability (callee_info->conds,
3666					    &e->predicate,
3667					    clause, es->param);
3668	  add_time = apply_probability ((gcov_type) add_time, prob);
3669	  if (add_time > MAX_TIME * INLINE_TIME_SCALE)
3670	    add_time = MAX_TIME * INLINE_TIME_SCALE;
3671	  if (prob != REG_BR_PROB_BASE
3672	      && dump_file && (dump_flags & TDF_DETAILS))
3673	    {
3674	      fprintf (dump_file, "\t\tScaling time by probability:%f\n",
3675		       (double) prob / REG_BR_PROB_BASE);
3676	    }
3677	  account_size_time (info, e->size, add_time, &p);
3678	}
3679    }
3680  remap_edge_summaries (edge, edge->callee, info, callee_info, operand_map,
3681			offset_map, clause, &toplev_predicate);
3682  remap_hint_predicate (info, callee_info,
3683			&callee_info->loop_iterations,
3684			operand_map, offset_map, clause, &toplev_predicate);
3685  remap_hint_predicate (info, callee_info,
3686			&callee_info->loop_stride,
3687			operand_map, offset_map, clause, &toplev_predicate);
3688  remap_hint_predicate (info, callee_info,
3689			&callee_info->array_index,
3690			operand_map, offset_map, clause, &toplev_predicate);
3691
3692  inline_update_callee_summaries (edge->callee,
3693				  inline_edge_summary (edge)->loop_depth);
3694
3695  /* We do not maintain predicates of inlined edges, free it.  */
3696  edge_set_predicate (edge, &true_p);
3697  /* Similarly remove param summaries.  */
3698  es->param.release ();
3699  operand_map.release ();
3700  offset_map.release ();
3701}
3702
3703/* For performance reasons inline_merge_summary is not updating overall size
3704   and time.  Recompute it.  */
3705
3706void
3707inline_update_overall_summary (struct cgraph_node *node)
3708{
3709  struct inline_summary *info = inline_summaries->get (node);
3710  size_time_entry *e;
3711  int i;
3712
3713  info->size = 0;
3714  info->time = 0;
3715  for (i = 0; vec_safe_iterate (info->entry, i, &e); i++)
3716    {
3717      info->size += e->size, info->time += e->time;
3718      if (info->time > MAX_TIME * INLINE_TIME_SCALE)
3719	info->time = MAX_TIME * INLINE_TIME_SCALE;
3720    }
3721  estimate_calls_size_and_time (node, &info->size, &info->min_size,
3722				&info->time, NULL,
3723				~(clause_t) (1 << predicate_false_condition),
3724				vNULL, vNULL, vNULL);
3725  info->time = (info->time + INLINE_TIME_SCALE / 2) / INLINE_TIME_SCALE;
3726  info->size = (info->size + INLINE_SIZE_SCALE / 2) / INLINE_SIZE_SCALE;
3727}
3728
3729/* Return hints derrived from EDGE.   */
3730int
3731simple_edge_hints (struct cgraph_edge *edge)
3732{
3733  int hints = 0;
3734  struct cgraph_node *to = (edge->caller->global.inlined_to
3735			    ? edge->caller->global.inlined_to : edge->caller);
3736  struct cgraph_node *callee = edge->callee->ultimate_alias_target ();
3737  if (inline_summaries->get (to)->scc_no
3738      && inline_summaries->get (to)->scc_no
3739	 == inline_summaries->get (callee)->scc_no
3740      && !edge->recursive_p ())
3741    hints |= INLINE_HINT_same_scc;
3742
3743  if (callee->lto_file_data && edge->caller->lto_file_data
3744      && edge->caller->lto_file_data != callee->lto_file_data
3745      && !callee->merged)
3746    hints |= INLINE_HINT_cross_module;
3747
3748  return hints;
3749}
3750
3751/* Estimate the time cost for the caller when inlining EDGE.
3752   Only to be called via estimate_edge_time, that handles the
3753   caching mechanism.
3754
3755   When caching, also update the cache entry.  Compute both time and
3756   size, since we always need both metrics eventually.  */
3757
3758int
3759do_estimate_edge_time (struct cgraph_edge *edge)
3760{
3761  int time;
3762  int size;
3763  inline_hints hints;
3764  struct cgraph_node *callee;
3765  clause_t clause;
3766  vec<tree> known_vals;
3767  vec<ipa_polymorphic_call_context> known_contexts;
3768  vec<ipa_agg_jump_function_p> known_aggs;
3769  struct inline_edge_summary *es = inline_edge_summary (edge);
3770  int min_size;
3771
3772  callee = edge->callee->ultimate_alias_target ();
3773
3774  gcc_checking_assert (edge->inline_failed);
3775  evaluate_properties_for_edge (edge, true,
3776				&clause, &known_vals, &known_contexts,
3777				&known_aggs);
3778  estimate_node_size_and_time (callee, clause, known_vals, known_contexts,
3779			       known_aggs, &size, &min_size, &time, &hints, es->param);
3780
3781  /* When we have profile feedback, we can quite safely identify hot
3782     edges and for those we disable size limits.  Don't do that when
3783     probability that caller will call the callee is low however, since it
3784     may hurt optimization of the caller's hot path.  */
3785  if (edge->count && edge->maybe_hot_p ()
3786      && (edge->count * 2
3787          > (edge->caller->global.inlined_to
3788	     ? edge->caller->global.inlined_to->count : edge->caller->count)))
3789    hints |= INLINE_HINT_known_hot;
3790
3791  known_vals.release ();
3792  known_contexts.release ();
3793  known_aggs.release ();
3794  gcc_checking_assert (size >= 0);
3795  gcc_checking_assert (time >= 0);
3796
3797  /* When caching, update the cache entry.  */
3798  if (edge_growth_cache.exists ())
3799    {
3800      inline_summaries->get (edge->callee)->min_size = min_size;
3801      if ((int) edge_growth_cache.length () <= edge->uid)
3802	edge_growth_cache.safe_grow_cleared (symtab->edges_max_uid);
3803      edge_growth_cache[edge->uid].time = time + (time >= 0);
3804
3805      edge_growth_cache[edge->uid].size = size + (size >= 0);
3806      hints |= simple_edge_hints (edge);
3807      edge_growth_cache[edge->uid].hints = hints + 1;
3808    }
3809  return time;
3810}
3811
3812
3813/* Return estimated callee growth after inlining EDGE.
3814   Only to be called via estimate_edge_size.  */
3815
3816int
3817do_estimate_edge_size (struct cgraph_edge *edge)
3818{
3819  int size;
3820  struct cgraph_node *callee;
3821  clause_t clause;
3822  vec<tree> known_vals;
3823  vec<ipa_polymorphic_call_context> known_contexts;
3824  vec<ipa_agg_jump_function_p> known_aggs;
3825
3826  /* When we do caching, use do_estimate_edge_time to populate the entry.  */
3827
3828  if (edge_growth_cache.exists ())
3829    {
3830      do_estimate_edge_time (edge);
3831      size = edge_growth_cache[edge->uid].size;
3832      gcc_checking_assert (size);
3833      return size - (size > 0);
3834    }
3835
3836  callee = edge->callee->ultimate_alias_target ();
3837
3838  /* Early inliner runs without caching, go ahead and do the dirty work.  */
3839  gcc_checking_assert (edge->inline_failed);
3840  evaluate_properties_for_edge (edge, true,
3841				&clause, &known_vals, &known_contexts,
3842				&known_aggs);
3843  estimate_node_size_and_time (callee, clause, known_vals, known_contexts,
3844			       known_aggs, &size, NULL, NULL, NULL, vNULL);
3845  known_vals.release ();
3846  known_contexts.release ();
3847  known_aggs.release ();
3848  return size;
3849}
3850
3851
3852/* Estimate the growth of the caller when inlining EDGE.
3853   Only to be called via estimate_edge_size.  */
3854
3855inline_hints
3856do_estimate_edge_hints (struct cgraph_edge *edge)
3857{
3858  inline_hints hints;
3859  struct cgraph_node *callee;
3860  clause_t clause;
3861  vec<tree> known_vals;
3862  vec<ipa_polymorphic_call_context> known_contexts;
3863  vec<ipa_agg_jump_function_p> known_aggs;
3864
3865  /* When we do caching, use do_estimate_edge_time to populate the entry.  */
3866
3867  if (edge_growth_cache.exists ())
3868    {
3869      do_estimate_edge_time (edge);
3870      hints = edge_growth_cache[edge->uid].hints;
3871      gcc_checking_assert (hints);
3872      return hints - 1;
3873    }
3874
3875  callee = edge->callee->ultimate_alias_target ();
3876
3877  /* Early inliner runs without caching, go ahead and do the dirty work.  */
3878  gcc_checking_assert (edge->inline_failed);
3879  evaluate_properties_for_edge (edge, true,
3880				&clause, &known_vals, &known_contexts,
3881				&known_aggs);
3882  estimate_node_size_and_time (callee, clause, known_vals, known_contexts,
3883			       known_aggs, NULL, NULL, NULL, &hints, vNULL);
3884  known_vals.release ();
3885  known_contexts.release ();
3886  known_aggs.release ();
3887  hints |= simple_edge_hints (edge);
3888  return hints;
3889}
3890
3891
3892/* Estimate self time of the function NODE after inlining EDGE.  */
3893
3894int
3895estimate_time_after_inlining (struct cgraph_node *node,
3896			      struct cgraph_edge *edge)
3897{
3898  struct inline_edge_summary *es = inline_edge_summary (edge);
3899  if (!es->predicate || !false_predicate_p (es->predicate))
3900    {
3901      gcov_type time =
3902	inline_summaries->get (node)->time + estimate_edge_time (edge);
3903      if (time < 0)
3904	time = 0;
3905      if (time > MAX_TIME)
3906	time = MAX_TIME;
3907      return time;
3908    }
3909  return inline_summaries->get (node)->time;
3910}
3911
3912
3913/* Estimate the size of NODE after inlining EDGE which should be an
3914   edge to either NODE or a call inlined into NODE.  */
3915
3916int
3917estimate_size_after_inlining (struct cgraph_node *node,
3918			      struct cgraph_edge *edge)
3919{
3920  struct inline_edge_summary *es = inline_edge_summary (edge);
3921  if (!es->predicate || !false_predicate_p (es->predicate))
3922    {
3923      int size = inline_summaries->get (node)->size + estimate_edge_growth (edge);
3924      gcc_assert (size >= 0);
3925      return size;
3926    }
3927  return inline_summaries->get (node)->size;
3928}
3929
3930
3931struct growth_data
3932{
3933  struct cgraph_node *node;
3934  bool self_recursive;
3935  bool uninlinable;
3936  int growth;
3937};
3938
3939
3940/* Worker for do_estimate_growth.  Collect growth for all callers.  */
3941
3942static bool
3943do_estimate_growth_1 (struct cgraph_node *node, void *data)
3944{
3945  struct cgraph_edge *e;
3946  struct growth_data *d = (struct growth_data *) data;
3947
3948  for (e = node->callers; e; e = e->next_caller)
3949    {
3950      gcc_checking_assert (e->inline_failed);
3951
3952      if (cgraph_inline_failed_type (e->inline_failed) == CIF_FINAL_ERROR)
3953	{
3954	  d->uninlinable = true;
3955          continue;
3956	}
3957
3958      if (e->recursive_p ())
3959	{
3960	  d->self_recursive = true;
3961	  continue;
3962	}
3963      d->growth += estimate_edge_growth (e);
3964    }
3965  return false;
3966}
3967
3968
3969/* Estimate the growth caused by inlining NODE into all callees.  */
3970
3971int
3972estimate_growth (struct cgraph_node *node)
3973{
3974  struct growth_data d = { node, false, false, 0 };
3975  struct inline_summary *info = inline_summaries->get (node);
3976
3977  node->call_for_symbol_and_aliases (do_estimate_growth_1, &d, true);
3978
3979  /* For self recursive functions the growth estimation really should be
3980     infinity.  We don't want to return very large values because the growth
3981     plays various roles in badness computation fractions.  Be sure to not
3982     return zero or negative growths. */
3983  if (d.self_recursive)
3984    d.growth = d.growth < info->size ? info->size : d.growth;
3985  else if (DECL_EXTERNAL (node->decl) || d.uninlinable)
3986    ;
3987  else
3988    {
3989      if (node->will_be_removed_from_program_if_no_direct_calls_p ())
3990	d.growth -= info->size;
3991      /* COMDAT functions are very often not shared across multiple units
3992         since they come from various template instantiations.
3993         Take this into account.  */
3994      else if (DECL_COMDAT (node->decl)
3995	       && node->can_remove_if_no_direct_calls_p ())
3996	d.growth -= (info->size
3997		     * (100 - PARAM_VALUE (PARAM_COMDAT_SHARING_PROBABILITY))
3998		     + 50) / 100;
3999    }
4000
4001  return d.growth;
4002}
4003
4004/* Verify if there are fewer than MAX_CALLERS.  */
4005
4006static bool
4007check_callers (cgraph_node *node, int *max_callers)
4008{
4009  ipa_ref *ref;
4010
4011  if (!node->can_remove_if_no_direct_calls_and_refs_p ())
4012    return true;
4013
4014  for (cgraph_edge *e = node->callers; e; e = e->next_caller)
4015    {
4016      (*max_callers)--;
4017      if (!*max_callers
4018	  || cgraph_inline_failed_type (e->inline_failed) == CIF_FINAL_ERROR)
4019	return true;
4020    }
4021
4022  FOR_EACH_ALIAS (node, ref)
4023    if (check_callers (dyn_cast <cgraph_node *> (ref->referring), max_callers))
4024      return true;
4025
4026  return false;
4027}
4028
4029
4030/* Make cheap estimation if growth of NODE is likely positive knowing
4031   EDGE_GROWTH of one particular edge.
4032   We assume that most of other edges will have similar growth
4033   and skip computation if there are too many callers.  */
4034
4035bool
4036growth_likely_positive (struct cgraph_node *node,
4037		        int edge_growth)
4038{
4039  int max_callers;
4040  struct cgraph_edge *e;
4041  gcc_checking_assert (edge_growth > 0);
4042
4043  /* First quickly check if NODE is removable at all.  */
4044  if (DECL_EXTERNAL (node->decl))
4045    return true;
4046  if (!node->can_remove_if_no_direct_calls_and_refs_p ()
4047      || node->address_taken)
4048    return true;
4049
4050  max_callers = inline_summaries->get (node)->size * 4 / edge_growth + 2;
4051
4052  for (e = node->callers; e; e = e->next_caller)
4053    {
4054      max_callers--;
4055      if (!max_callers
4056	  || cgraph_inline_failed_type (e->inline_failed) == CIF_FINAL_ERROR)
4057	return true;
4058    }
4059
4060  ipa_ref *ref;
4061  FOR_EACH_ALIAS (node, ref)
4062    if (check_callers (dyn_cast <cgraph_node *> (ref->referring), &max_callers))
4063      return true;
4064
4065  /* Unlike for functions called once, we play unsafe with
4066     COMDATs.  We can allow that since we know functions
4067     in consideration are small (and thus risk is small) and
4068     moreover grow estimates already accounts that COMDAT
4069     functions may or may not disappear when eliminated from
4070     current unit. With good probability making aggressive
4071     choice in all units is going to make overall program
4072     smaller.  */
4073  if (DECL_COMDAT (node->decl))
4074    {
4075      if (!node->can_remove_if_no_direct_calls_p ())
4076	return true;
4077    }
4078  else if (!node->will_be_removed_from_program_if_no_direct_calls_p ())
4079    return true;
4080
4081  return estimate_growth (node) > 0;
4082}
4083
4084
4085/* This function performs intraprocedural analysis in NODE that is required to
4086   inline indirect calls.  */
4087
4088static void
4089inline_indirect_intraprocedural_analysis (struct cgraph_node *node)
4090{
4091  ipa_analyze_node (node);
4092  if (dump_file && (dump_flags & TDF_DETAILS))
4093    {
4094      ipa_print_node_params (dump_file, node);
4095      ipa_print_node_jump_functions (dump_file, node);
4096    }
4097}
4098
4099
4100/* Note function body size.  */
4101
4102void
4103inline_analyze_function (struct cgraph_node *node)
4104{
4105  push_cfun (DECL_STRUCT_FUNCTION (node->decl));
4106
4107  if (dump_file)
4108    fprintf (dump_file, "\nAnalyzing function: %s/%u\n",
4109	     node->name (), node->order);
4110  if (opt_for_fn (node->decl, optimize) && !node->thunk.thunk_p)
4111    inline_indirect_intraprocedural_analysis (node);
4112  compute_inline_parameters (node, false);
4113  if (!optimize)
4114    {
4115      struct cgraph_edge *e;
4116      for (e = node->callees; e; e = e->next_callee)
4117	{
4118	  if (e->inline_failed == CIF_FUNCTION_NOT_CONSIDERED)
4119	    e->inline_failed = CIF_FUNCTION_NOT_OPTIMIZED;
4120	  e->call_stmt_cannot_inline_p = true;
4121	}
4122      for (e = node->indirect_calls; e; e = e->next_callee)
4123	{
4124	  if (e->inline_failed == CIF_FUNCTION_NOT_CONSIDERED)
4125	    e->inline_failed = CIF_FUNCTION_NOT_OPTIMIZED;
4126	  e->call_stmt_cannot_inline_p = true;
4127	}
4128    }
4129
4130  pop_cfun ();
4131}
4132
4133
4134/* Called when new function is inserted to callgraph late.  */
4135
4136void
4137inline_summary_t::insert (struct cgraph_node *node, inline_summary *)
4138{
4139  inline_analyze_function (node);
4140}
4141
4142/* Note function body size.  */
4143
4144void
4145inline_generate_summary (void)
4146{
4147  struct cgraph_node *node;
4148
4149  /* When not optimizing, do not bother to analyze.  Inlining is still done
4150     because edge redirection needs to happen there.  */
4151  if (!optimize && !flag_generate_lto && !flag_generate_offload && !flag_wpa)
4152    return;
4153
4154  if (!inline_summaries)
4155    inline_summaries = (inline_summary_t*) inline_summary_t::create_ggc (symtab);
4156
4157  inline_summaries->enable_insertion_hook ();
4158
4159  ipa_register_cgraph_hooks ();
4160  inline_free_summary ();
4161
4162  FOR_EACH_DEFINED_FUNCTION (node)
4163    if (!node->alias)
4164      inline_analyze_function (node);
4165}
4166
4167
4168/* Read predicate from IB.  */
4169
4170static struct predicate
4171read_predicate (struct lto_input_block *ib)
4172{
4173  struct predicate out;
4174  clause_t clause;
4175  int k = 0;
4176
4177  do
4178    {
4179      gcc_assert (k <= MAX_CLAUSES);
4180      clause = out.clause[k++] = streamer_read_uhwi (ib);
4181    }
4182  while (clause);
4183
4184  /* Zero-initialize the remaining clauses in OUT.  */
4185  while (k <= MAX_CLAUSES)
4186    out.clause[k++] = 0;
4187
4188  return out;
4189}
4190
4191
4192/* Write inline summary for edge E to OB.  */
4193
4194static void
4195read_inline_edge_summary (struct lto_input_block *ib, struct cgraph_edge *e)
4196{
4197  struct inline_edge_summary *es = inline_edge_summary (e);
4198  struct predicate p;
4199  int length, i;
4200
4201  es->call_stmt_size = streamer_read_uhwi (ib);
4202  es->call_stmt_time = streamer_read_uhwi (ib);
4203  es->loop_depth = streamer_read_uhwi (ib);
4204  p = read_predicate (ib);
4205  edge_set_predicate (e, &p);
4206  length = streamer_read_uhwi (ib);
4207  if (length)
4208    {
4209      es->param.safe_grow_cleared (length);
4210      for (i = 0; i < length; i++)
4211	es->param[i].change_prob = streamer_read_uhwi (ib);
4212    }
4213}
4214
4215
4216/* Stream in inline summaries from the section.  */
4217
4218static void
4219inline_read_section (struct lto_file_decl_data *file_data, const char *data,
4220		     size_t len)
4221{
4222  const struct lto_function_header *header =
4223    (const struct lto_function_header *) data;
4224  const int cfg_offset = sizeof (struct lto_function_header);
4225  const int main_offset = cfg_offset + header->cfg_size;
4226  const int string_offset = main_offset + header->main_size;
4227  struct data_in *data_in;
4228  unsigned int i, count2, j;
4229  unsigned int f_count;
4230
4231  lto_input_block ib ((const char *) data + main_offset, header->main_size,
4232		      file_data->mode_table);
4233
4234  data_in =
4235    lto_data_in_create (file_data, (const char *) data + string_offset,
4236			header->string_size, vNULL);
4237  f_count = streamer_read_uhwi (&ib);
4238  for (i = 0; i < f_count; i++)
4239    {
4240      unsigned int index;
4241      struct cgraph_node *node;
4242      struct inline_summary *info;
4243      lto_symtab_encoder_t encoder;
4244      struct bitpack_d bp;
4245      struct cgraph_edge *e;
4246      predicate p;
4247
4248      index = streamer_read_uhwi (&ib);
4249      encoder = file_data->symtab_node_encoder;
4250      node = dyn_cast<cgraph_node *> (lto_symtab_encoder_deref (encoder,
4251								index));
4252      info = inline_summaries->get (node);
4253
4254      info->estimated_stack_size
4255	= info->estimated_self_stack_size = streamer_read_uhwi (&ib);
4256      info->size = info->self_size = streamer_read_uhwi (&ib);
4257      info->time = info->self_time = streamer_read_uhwi (&ib);
4258
4259      bp = streamer_read_bitpack (&ib);
4260      info->inlinable = bp_unpack_value (&bp, 1);
4261      info->contains_cilk_spawn = bp_unpack_value (&bp, 1);
4262
4263      count2 = streamer_read_uhwi (&ib);
4264      gcc_assert (!info->conds);
4265      for (j = 0; j < count2; j++)
4266	{
4267	  struct condition c;
4268	  c.operand_num = streamer_read_uhwi (&ib);
4269	  c.size = streamer_read_uhwi (&ib);
4270	  c.code = (enum tree_code) streamer_read_uhwi (&ib);
4271	  c.val = stream_read_tree (&ib, data_in);
4272	  bp = streamer_read_bitpack (&ib);
4273	  c.agg_contents = bp_unpack_value (&bp, 1);
4274	  c.by_ref = bp_unpack_value (&bp, 1);
4275	  if (c.agg_contents)
4276	    c.offset = streamer_read_uhwi (&ib);
4277	  vec_safe_push (info->conds, c);
4278	}
4279      count2 = streamer_read_uhwi (&ib);
4280      gcc_assert (!info->entry);
4281      for (j = 0; j < count2; j++)
4282	{
4283	  struct size_time_entry e;
4284
4285	  e.size = streamer_read_uhwi (&ib);
4286	  e.time = streamer_read_uhwi (&ib);
4287	  e.predicate = read_predicate (&ib);
4288
4289	  vec_safe_push (info->entry, e);
4290	}
4291
4292      p = read_predicate (&ib);
4293      set_hint_predicate (&info->loop_iterations, p);
4294      p = read_predicate (&ib);
4295      set_hint_predicate (&info->loop_stride, p);
4296      p = read_predicate (&ib);
4297      set_hint_predicate (&info->array_index, p);
4298      for (e = node->callees; e; e = e->next_callee)
4299	read_inline_edge_summary (&ib, e);
4300      for (e = node->indirect_calls; e; e = e->next_callee)
4301	read_inline_edge_summary (&ib, e);
4302    }
4303
4304  lto_free_section_data (file_data, LTO_section_inline_summary, NULL, data,
4305			 len);
4306  lto_data_in_delete (data_in);
4307}
4308
4309
4310/* Read inline summary.  Jump functions are shared among ipa-cp
4311   and inliner, so when ipa-cp is active, we don't need to write them
4312   twice.  */
4313
4314void
4315inline_read_summary (void)
4316{
4317  struct lto_file_decl_data **file_data_vec = lto_get_file_decl_data ();
4318  struct lto_file_decl_data *file_data;
4319  unsigned int j = 0;
4320
4321  inline_summary_alloc ();
4322
4323  while ((file_data = file_data_vec[j++]))
4324    {
4325      size_t len;
4326      const char *data = lto_get_section_data (file_data,
4327					       LTO_section_inline_summary,
4328					       NULL, &len);
4329      if (data)
4330	inline_read_section (file_data, data, len);
4331      else
4332	/* Fatal error here.  We do not want to support compiling ltrans units
4333	   with different version of compiler or different flags than the WPA
4334	   unit, so this should never happen.  */
4335	fatal_error (input_location,
4336		     "ipa inline summary is missing in input file");
4337    }
4338  if (optimize)
4339    {
4340      ipa_register_cgraph_hooks ();
4341      if (!flag_ipa_cp)
4342	ipa_prop_read_jump_functions ();
4343    }
4344
4345  gcc_assert (inline_summaries);
4346  inline_summaries->enable_insertion_hook ();
4347}
4348
4349
4350/* Write predicate P to OB.  */
4351
4352static void
4353write_predicate (struct output_block *ob, struct predicate *p)
4354{
4355  int j;
4356  if (p)
4357    for (j = 0; p->clause[j]; j++)
4358      {
4359	gcc_assert (j < MAX_CLAUSES);
4360	streamer_write_uhwi (ob, p->clause[j]);
4361      }
4362  streamer_write_uhwi (ob, 0);
4363}
4364
4365
4366/* Write inline summary for edge E to OB.  */
4367
4368static void
4369write_inline_edge_summary (struct output_block *ob, struct cgraph_edge *e)
4370{
4371  struct inline_edge_summary *es = inline_edge_summary (e);
4372  int i;
4373
4374  streamer_write_uhwi (ob, es->call_stmt_size);
4375  streamer_write_uhwi (ob, es->call_stmt_time);
4376  streamer_write_uhwi (ob, es->loop_depth);
4377  write_predicate (ob, es->predicate);
4378  streamer_write_uhwi (ob, es->param.length ());
4379  for (i = 0; i < (int) es->param.length (); i++)
4380    streamer_write_uhwi (ob, es->param[i].change_prob);
4381}
4382
4383
4384/* Write inline summary for node in SET.
4385   Jump functions are shared among ipa-cp and inliner, so when ipa-cp is
4386   active, we don't need to write them twice.  */
4387
4388void
4389inline_write_summary (void)
4390{
4391  struct cgraph_node *node;
4392  struct output_block *ob = create_output_block (LTO_section_inline_summary);
4393  lto_symtab_encoder_t encoder = ob->decl_state->symtab_node_encoder;
4394  unsigned int count = 0;
4395  int i;
4396
4397  for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
4398    {
4399      symtab_node *snode = lto_symtab_encoder_deref (encoder, i);
4400      cgraph_node *cnode = dyn_cast <cgraph_node *> (snode);
4401      if (cnode && cnode->definition && !cnode->alias)
4402	count++;
4403    }
4404  streamer_write_uhwi (ob, count);
4405
4406  for (i = 0; i < lto_symtab_encoder_size (encoder); i++)
4407    {
4408      symtab_node *snode = lto_symtab_encoder_deref (encoder, i);
4409      cgraph_node *cnode = dyn_cast <cgraph_node *> (snode);
4410      if (cnode && (node = cnode)->definition && !node->alias)
4411	{
4412	  struct inline_summary *info = inline_summaries->get (node);
4413	  struct bitpack_d bp;
4414	  struct cgraph_edge *edge;
4415	  int i;
4416	  size_time_entry *e;
4417	  struct condition *c;
4418
4419	  streamer_write_uhwi (ob,
4420			       lto_symtab_encoder_encode (encoder,
4421
4422							  node));
4423	  streamer_write_hwi (ob, info->estimated_self_stack_size);
4424	  streamer_write_hwi (ob, info->self_size);
4425	  streamer_write_hwi (ob, info->self_time);
4426	  bp = bitpack_create (ob->main_stream);
4427	  bp_pack_value (&bp, info->inlinable, 1);
4428	  bp_pack_value (&bp, info->contains_cilk_spawn, 1);
4429	  streamer_write_bitpack (&bp);
4430	  streamer_write_uhwi (ob, vec_safe_length (info->conds));
4431	  for (i = 0; vec_safe_iterate (info->conds, i, &c); i++)
4432	    {
4433	      streamer_write_uhwi (ob, c->operand_num);
4434	      streamer_write_uhwi (ob, c->size);
4435	      streamer_write_uhwi (ob, c->code);
4436	      stream_write_tree (ob, c->val, true);
4437	      bp = bitpack_create (ob->main_stream);
4438	      bp_pack_value (&bp, c->agg_contents, 1);
4439	      bp_pack_value (&bp, c->by_ref, 1);
4440	      streamer_write_bitpack (&bp);
4441	      if (c->agg_contents)
4442		streamer_write_uhwi (ob, c->offset);
4443	    }
4444	  streamer_write_uhwi (ob, vec_safe_length (info->entry));
4445	  for (i = 0; vec_safe_iterate (info->entry, i, &e); i++)
4446	    {
4447	      streamer_write_uhwi (ob, e->size);
4448	      streamer_write_uhwi (ob, e->time);
4449	      write_predicate (ob, &e->predicate);
4450	    }
4451	  write_predicate (ob, info->loop_iterations);
4452	  write_predicate (ob, info->loop_stride);
4453	  write_predicate (ob, info->array_index);
4454	  for (edge = node->callees; edge; edge = edge->next_callee)
4455	    write_inline_edge_summary (ob, edge);
4456	  for (edge = node->indirect_calls; edge; edge = edge->next_callee)
4457	    write_inline_edge_summary (ob, edge);
4458	}
4459    }
4460  streamer_write_char_stream (ob->main_stream, 0);
4461  produce_asm (ob, NULL);
4462  destroy_output_block (ob);
4463
4464  if (optimize && !flag_ipa_cp)
4465    ipa_prop_write_jump_functions ();
4466}
4467
4468
4469/* Release inline summary.  */
4470
4471void
4472inline_free_summary (void)
4473{
4474  struct cgraph_node *node;
4475  if (edge_removal_hook_holder)
4476    symtab->remove_edge_removal_hook (edge_removal_hook_holder);
4477  edge_removal_hook_holder = NULL;
4478  if (edge_duplication_hook_holder)
4479    symtab->remove_edge_duplication_hook (edge_duplication_hook_holder);
4480  edge_duplication_hook_holder = NULL;
4481  if (!inline_edge_summary_vec.exists ())
4482    return;
4483  FOR_EACH_DEFINED_FUNCTION (node)
4484    if (!node->alias)
4485      reset_inline_summary (node, inline_summaries->get (node));
4486  inline_summaries->release ();
4487  inline_summaries = NULL;
4488  inline_edge_summary_vec.release ();
4489  if (edge_predicate_pool)
4490    free_alloc_pool (edge_predicate_pool);
4491  edge_predicate_pool = 0;
4492}
4493