diff mbox series

Fix PR90450

Message ID alpine.LSU.2.20.1905220943010.10704@zhemvz.fhfr.qr
State New
Headers show
Series Fix PR90450 | expand

Commit Message

Richard Biener May 22, 2019, 7:43 a.m. UTC
The hash-table checker discovered a bug in LIMs new fancy hash/compare
logic, fixed as follows.

Bootstrapped and tested on x86_64-unknown-linux-gnu, applied to trunk,
queued for branch.

Richard.

2019-05-22  Richard Biener  <rguenther@suse.de>

	PR tree-optimization/90450
	* tree-ssa-loop-im.c (struct im_mem_ref): Add ref_decomposed.
	(mem_ref_hasher::equal): Check it.
	(mem_ref_alloc): Initialize it.
	(gather_mem_refs_stmt): Set it.
diff mbox series

Patch

Index: gcc/tree-ssa-loop-im.c
===================================================================
--- gcc/tree-ssa-loop-im.c	(revision 271415)
+++ gcc/tree-ssa-loop-im.c	(working copy)
@@ -115,9 +115,10 @@  struct mem_ref_loc
 
 struct im_mem_ref
 {
-  unsigned id : 31;		/* ID assigned to the memory reference
+  unsigned id : 30;		/* ID assigned to the memory reference
 				   (its index in memory_accesses.refs_list)  */
   unsigned ref_canonical : 1;   /* Whether mem.ref was canonicalized.  */
+  unsigned ref_decomposed : 1;  /* Whether the ref was hashed from mem.  */
   hashval_t hash;		/* Its hash value.  */
 
   /* The memory access itself and associated caching of alias-oracle
@@ -173,7 +174,8 @@  inline bool
 mem_ref_hasher::equal (const im_mem_ref *mem1, const ao_ref *obj2)
 {
   if (obj2->max_size_known_p ())
-    return (operand_equal_p (mem1->mem.base, obj2->base, 0)
+    return (mem1->ref_decomposed
+	    && operand_equal_p (mem1->mem.base, obj2->base, 0)
 	    && known_eq (mem1->mem.offset, obj2->offset)
 	    && known_eq (mem1->mem.size, obj2->size)
 	    && known_eq (mem1->mem.max_size, obj2->max_size)
@@ -1389,6 +1391,7 @@  mem_ref_alloc (ao_ref *mem, unsigned has
     ao_ref_init (&ref->mem, error_mark_node);
   ref->id = id;
   ref->ref_canonical = false;
+  ref->ref_decomposed = false;
   ref->hash = hash;
   ref->stored = NULL;
   bitmap_initialize (&ref->indep_loop, &lim_bitmap_obstack);
@@ -1476,6 +1479,7 @@  gather_mem_refs_stmt (struct loop *loop,
       HOST_WIDE_INT offset, size, max_size;
       poly_int64 saved_maxsize = aor.max_size, mem_off;
       tree mem_base;
+      bool ref_decomposed;
       if (aor.max_size_known_p ()
 	  && aor.offset.is_constant (&offset)
 	  && aor.size.is_constant (&size)
@@ -1489,12 +1493,14 @@  gather_mem_refs_stmt (struct loop *loop,
 		       aor.size)
 	  && (mem_base = get_addr_base_and_unit_offset (aor.ref, &mem_off)))
 	{
+	  ref_decomposed = true;
 	  hash = iterative_hash_expr (ao_ref_base (&aor), 0);
 	  hash = iterative_hash_host_wide_int (offset, hash);
 	  hash = iterative_hash_host_wide_int (size, hash);
 	}
       else
 	{
+	  ref_decomposed = false;
 	  hash = iterative_hash_expr (aor.ref, 0);
 	  aor.max_size = -1;
 	}
@@ -1543,6 +1549,7 @@  gather_mem_refs_stmt (struct loop *loop,
 	{
 	  id = memory_accesses.refs_list.length ();
 	  ref = mem_ref_alloc (&aor, hash, id);
+	  ref->ref_decomposed = ref_decomposed;
 	  memory_accesses.refs_list.safe_push (ref);
 	  *slot = ref;