diff mbox series

Enable early inlining into always_inline functions

Message ID ZJyhfQRGFnzqdjFl@kam.mff.cuni.cz
State New
Headers show
Series Enable early inlining into always_inline functions | expand

Commit Message

Jan Hubicka June 28, 2023, 9:09 p.m. UTC
Hi,
early inliner currently skips always_inline functions and moreover we ignore
calls from always_inline in ipa_reverse_postorder.  This leads to disabling
most of propagation done using early optimization that is quite bad when
early inline functions are not leaf functions, which is now quite common
in libstdc++.

This patch instead of fully disabling the inline checks calls in callee.
I am quite conservative about what can be inlined as this patch is bit
touchy anyway.  To avoid problems with always_inline being optimized
after early inline I extended inline_always_inline_functions to lazilly
compute fnsummary when needed.

Bootstrapped/regtested x86_64-linux, will commit it shortly.

gcc/ChangeLog:

	PR middle-end/110334
	* ipa-fnsummary.h (ipa_fn_summary): Add
	safe_to_inline_to_always_inline.
	* ipa-inline.cc (can_early_inline_edge_p): ICE
	if SSA is not built; do cycle checking for
	always_inline functions.
	(inline_always_inline_functions): Be recrusive;
	watch for cycles; do not updat overall summary.
	(early_inliner): Do not give up on always_inlines.
	* ipa-utils.cc (ipa_reverse_postorder): Do not skip
	always inlines.

gcc/testsuite/ChangeLog:

	PR middle-end/110334
	* g++.dg/opt/pr66119.C: Disable early inlining.
	* gcc.c-torture/compile/pr110334.c: New test.
	* gcc.dg/tree-ssa/pr110334.c: New test.
diff mbox series

Patch

diff --git a/gcc/ipa-fnsummary.h b/gcc/ipa-fnsummary.h
index fcc01167d0d..0c5a81e2dca 100644
--- a/gcc/ipa-fnsummary.h
+++ b/gcc/ipa-fnsummary.h
@@ -126,8 +126,8 @@  public:
   ipa_fn_summary ()
     : min_size (0),
       inlinable (false), single_caller (false),
-      fp_expressions (false), target_info (0),
-      estimated_stack_size (false),
+      fp_expressions (false), safe_to_inline_to_always_inline (0),
+      target_info (0), estimated_stack_size (false),
       time (0), conds (NULL),
       size_time_table (), call_size_time_table (vNULL),
       loop_iterations (NULL), loop_strides (NULL),
@@ -165,6 +165,8 @@  public:
   unsigned int single_caller : 1;
   /* True if function contains any floating point expressions.  */
   unsigned int fp_expressions : 1;
+  /* Cache for analysis of can_early_inline_edge_p.  */
+  unsigned int safe_to_inline_to_always_inline : 2;
   /* Like fp_expressions field above, but it's to hold some target specific
      information, such as some target specific isa flags.  Note that for
      offloading target compilers, this field isn't streamed.  */
diff --git a/gcc/ipa-inline.cc b/gcc/ipa-inline.cc
index efc8df7d4e0..71a1c6ca68e 100644
--- a/gcc/ipa-inline.cc
+++ b/gcc/ipa-inline.cc
@@ -680,28 +680,60 @@  can_early_inline_edge_p (struct cgraph_edge *e)
       e->inline_failed = CIF_BODY_NOT_AVAILABLE;
       return false;
     }
-  /* In early inliner some of callees may not be in SSA form yet
-     (i.e. the callgraph is cyclic and we did not process
-     the callee by early inliner, yet).  We don't have CIF code for this
-     case; later we will re-do the decision in the real inliner.  */
-  if (!gimple_in_ssa_p (DECL_STRUCT_FUNCTION (e->caller->decl))
-      || !gimple_in_ssa_p (DECL_STRUCT_FUNCTION (callee->decl)))
-    {
-      if (dump_enabled_p ())
-	dump_printf_loc (MSG_MISSED_OPTIMIZATION, e->call_stmt,
-			 "  edge not inlinable: not in SSA form\n");
-      return false;
-    }
-  else if (profile_arc_flag
-	   && ((lookup_attribute ("no_profile_instrument_function",
-				 DECL_ATTRIBUTES (caller->decl)) == NULL_TREE)
-	       != (lookup_attribute ("no_profile_instrument_function",
-				     DECL_ATTRIBUTES (callee->decl)) == NULL_TREE)))
+  gcc_assert (gimple_in_ssa_p (DECL_STRUCT_FUNCTION (e->caller->decl))
+	      && gimple_in_ssa_p (DECL_STRUCT_FUNCTION (callee->decl)));
+  if (profile_arc_flag
+      && ((lookup_attribute ("no_profile_instrument_function",
+			    DECL_ATTRIBUTES (caller->decl)) == NULL_TREE)
+	  != (lookup_attribute ("no_profile_instrument_function",
+				DECL_ATTRIBUTES (callee->decl)) == NULL_TREE)))
     return false;
 
   if (!can_inline_edge_p (e, true, true)
       || !can_inline_edge_by_limits_p (e, true, false, true))
     return false;
+  /* When inlining regular functions into always-inline functions
+     during early inlining watch for possible inline cycles.  */
+  if (DECL_DISREGARD_INLINE_LIMITS (caller->decl)
+      && lookup_attribute ("always_inline", DECL_ATTRIBUTES (caller->decl))
+      && (!DECL_DISREGARD_INLINE_LIMITS (callee->decl)
+	  || !lookup_attribute ("always_inline", DECL_ATTRIBUTES (callee->decl))))
+    {
+      /* If there are indirect calls, inlining may produce direct call.
+	 TODO: We may lift this restriction if we avoid errors on formely
+	 indirect calls to always_inline functions.  Taking address
+	 of always_inline function is generally bad idea and should
+	 have been declared as undefined, but sadly we allow this.  */
+      if (caller->indirect_calls || e->callee->indirect_calls)
+	return false;
+      ipa_fn_summary *callee_info = ipa_fn_summaries->get (callee);
+      if (callee_info->safe_to_inline_to_always_inline)
+	return callee_info->safe_to_inline_to_always_inline - 1;
+      for (cgraph_edge *e2 = callee->callees; e2; e2 = e2->next_callee)
+	{
+	  struct cgraph_node *callee2 = e2->callee->ultimate_alias_target ();
+	  /* As early inliner runs in RPO order, we will see uninlined
+	     always_inline calls only in the case of cyclic graphs.  */
+	  if (DECL_DISREGARD_INLINE_LIMITS (callee2->decl)
+	      || lookup_attribute ("always_inline", DECL_ATTRIBUTES (callee2->decl)))
+	    {
+	      callee_info->safe_to_inline_to_always_inline = 1;
+	      return false;
+	    }
+	  /* With LTO watch for case where function is later replaced
+	     by always_inline definition.
+	     TODO: We may either stop treating noninlined cross-module always
+	     inlines as errors, or we can extend decl merging to produce
+	     syntacic alias and honor always inline only in units it has
+	     been declared as such.  */
+	  if (flag_lto && callee2->externally_visible)
+	    {
+	      callee_info->safe_to_inline_to_always_inline = 1;
+	      return false;
+	    }
+	}
+      callee_info->safe_to_inline_to_always_inline = 2;
+    }
   return true;
 }
 
@@ -2896,18 +2928,21 @@  ipa_inline (void)
   return remove_functions ? TODO_remove_functions : 0;
 }
 
-/* Inline always-inline function calls in NODE.  */
+/* Inline always-inline function calls in NODE
+   (which itself is possibly inline).  */
 
 static bool
 inline_always_inline_functions (struct cgraph_node *node)
 {
   struct cgraph_edge *e;
   bool inlined = false;
 
   for (e = node->callees; e; e = e->next_callee)
     {
       struct cgraph_node *callee = e->callee->ultimate_alias_target ();
-      if (!DECL_DISREGARD_INLINE_LIMITS (callee->decl))
+      gcc_checking_assert (!callee->aux || callee->aux == (void *)(size_t)1);
+      if (!DECL_DISREGARD_INLINE_LIMITS (callee->decl)
+	  /* Watch for self-recursive cycles.  */
+	  || callee->aux)
 	continue;
 
       if (e->recursive_p ())
@@ -2919,6 +2954,9 @@  inline_always_inline_functions (struct cgraph_node *node)
 	  e->inline_failed = CIF_RECURSIVE_INLINING;
 	  continue;
 	}
+      if (callee->definition
+	  && !ipa_fn_summaries->get (callee))
+	compute_fn_summary (callee, true);
 
       if (!can_early_inline_edge_p (e))
 	{
@@ -2936,11 +2974,13 @@  inline_always_inline_functions (struct cgraph_node *node)
 			 "  Inlining %C into %C (always_inline).\n",
 			 e->callee, e->caller);
       inline_call (e, true, NULL, NULL, false);
+      callee->aux = (void *)(size_t)1;
+      /* Inline recursively to handle the case where always_inline function was
+	 not optimized yet since it is a part of a cycle in callgraph.  */
+      inline_always_inline_functions (e->callee);
+      callee->aux = NULL;
       inlined = true;
     }
-  if (inlined)
-    ipa_update_overall_fn_summary (node);
-
   return inlined;
 }
 
@@ -3034,18 +3074,7 @@  early_inliner (function *fun)
 
   if (!optimize
       || flag_no_inline
-      || !flag_early_inlining
-      /* Never inline regular functions into always-inline functions
-	 during incremental inlining.  This sucks as functions calling
-	 always inline functions will get less optimized, but at the
-	 same time inlining of functions calling always inline
-	 function into an always inline function might introduce
-	 cycles of edges to be always inlined in the callgraph.
-
-	 We might want to be smarter and just avoid this type of inlining.  */
-      || (DECL_DISREGARD_INLINE_LIMITS (node->decl)
-	  && lookup_attribute ("always_inline",
-			       DECL_ATTRIBUTES (node->decl))))
+      || !flag_early_inlining)
     ;
   else if (lookup_attribute ("flatten",
 			     DECL_ATTRIBUTES (node->decl)) != NULL)
@@ -3063,7 +3092,8 @@  early_inliner (function *fun)
       /* If some always_inline functions was inlined, apply the changes.
 	 This way we will not account always inline into growth limits and
 	 moreover we will inline calls from always inlines that we skipped
-	 previously because of conditional above.  */
+	 previously because of conditional in can_early_inline_edge_p
+	 which prevents some inlining to always_inline.  */
       if (inlined)
 	{
 	  timevar_push (TV_INTEGRATION);
diff --git a/gcc/ipa-utils.cc b/gcc/ipa-utils.cc
index 8badcc2c110..ce84d4712a3 100644
--- a/gcc/ipa-utils.cc
+++ b/gcc/ipa-utils.cc
@@ -323,13 +325,6 @@  ipa_reverse_postorder (struct cgraph_node **order)
 		      edge = stack[stack_size].edge;
 		      node2 = edge->caller;
 		      stack[stack_size].edge = edge->next_caller;
-		      /* Break possible cycles involving always-inline
-			 functions by ignoring edges from always-inline
-			 functions to non-always-inline functions.  */
-		      if (DECL_DISREGARD_INLINE_LIMITS (edge->caller->decl)
-			  && !DECL_DISREGARD_INLINE_LIMITS
-			    (edge->callee->function_symbol ()->decl))
-			node2 = NULL;
 		    }
 		  for (; stack[stack_size].node->iterate_referring (
 						       stack[stack_size].ref,
diff --git a/gcc/testsuite/g++.dg/opt/pr66119.C b/gcc/testsuite/g++.dg/opt/pr66119.C
index 5b420c23de8..c129633894f 100644
--- a/gcc/testsuite/g++.dg/opt/pr66119.C
+++ b/gcc/testsuite/g++.dg/opt/pr66119.C
@@ -3,7 +3,7 @@ 
    the value of MOVE_RATIO now is.  */
 
 /* { dg-do compile  { target { { i?86-*-* x86_64-*-* } && c++11 } }  }  */
-/* { dg-options "-O3 -mavx -fdump-tree-sra -march=slm -mtune=slm" } */
+/* { dg-options "-O3 -mavx -fdump-tree-sra -march=slm -mtune=slm -fno-early-inlining" } */
 
 #include <immintrin.h>
 
diff --git a/gcc/testsuite/gcc.c-torture/compile/pr110334.c b/gcc/testsuite/gcc.c-torture/compile/pr110334.c
new file mode 100644
index 00000000000..5524bf6aaf0
--- /dev/null
+++ b/gcc/testsuite/gcc.c-torture/compile/pr110334.c
@@ -0,0 +1,20 @@ 
+int n;
+typedef void (*fnptr)();
+fnptr get_me();
+__attribute__ ((always_inline))
+inline void test(void)
+{
+        if (n < 10)
+          (get_me())();
+        n++;
+        return;
+}
+fnptr get_me()
+{
+        return test;
+}
+void
+foo()
+{
+        test();
+}
diff --git a/gcc/testsuite/gcc.dg/tree-ssa/pr110334.c b/gcc/testsuite/gcc.dg/tree-ssa/pr110334.c
new file mode 100644
index 00000000000..bc085d4b169
--- /dev/null
+++ b/gcc/testsuite/gcc.dg/tree-ssa/pr110334.c
@@ -0,0 +1,17 @@ 
+/* { dg-do compile } */
+/* { dg-options "-O2 -fdump-tree-release_ssa" } */
+int a;
+int ret1()
+{
+	return a;
+}
+int inline
+__attribute__((always_inline)) aret1()
+{
+	return ret1();
+}
+int test()
+{
+  return aret1();
+}
+/* { dg-final { scan-tree-dump-not "= ret1" "release_ssa" } }  */