Message ID | CAKxPW65RS5caBn49H=BuJ+tKXhDJMmVSpor4ykCRokxrTNOTjw@mail.gmail.com |
---|---|
State | New |
Headers | show |
On Wed, Oct 24, 2012 at 2:53 AM, Sharad Singhai <singhai@google.com> wrote: > Okay, I have updated the patch (attached) to allow inlining of > dump_enabled_p. I needed to make alt_dump_file extern to for this > (dump_file was already extern). Also, I moved dump_file definition > from toplev.c to dumpfile.c, where it seems to logically belong. > > Bootstrapped and tested on x86_64. Okay for trunk? Ok! Thanks for keeping up with our changing minds ;) Richard. > Thanks, > Sharad > > 2012-10-24 Sharad Singhai <singhai@google.com> > > * dumpfile.c (dump_enabled_p): Make it inline and move the definition > to dumpfile.h. > (dump_kind_p): Deleted. Functionality replaced by dump_enabled_p. > Make alt_dump_file extern. > * dumpfile.h (dump_enabled_p): Move inline definition here. > (dump_kind_p): Delete declaration. > Add extern declaration of alt_dump_file. > * toplev.c: Move dump_file and dump_file_name to dumpfile.c. > * tree-vect-loop-manip.c: Replace all uses of dump_kind_p with > dump_enabled_p. > * tree-vectorizer.c: Likewise. > * tree-vect-loop.c: Likewise. > * tree-vect-data-refs.c: Likewise. > * tree-vect-patterns.c: Likewise. > * tree-vect-stmts.c: Likewise. > * tree-vect-slp.c: Likewise. > > Index: dumpfile.c > =================================================================== > --- dumpfile.c (revision 192695) > +++ dumpfile.c (working copy) > @@ -32,12 +32,18 @@ along with GCC; see the file COPYING3. If not see > > static int pflags; /* current dump_flags */ > static int alt_flags; /* current opt_info flags */ > -static FILE *alt_dump_file = NULL; > > static void dump_loc (int, FILE *, source_location); > static int dump_phase_enabled_p (int); > static FILE *dump_open_alternate_stream (struct dump_file_info *); > > +/* These are currently used for communicating between passes. > + However, instead of accessing them directly, the passes can use > + dump_printf () for dumps. */ > +FILE *dump_file = NULL; > +FILE *alt_dump_file = NULL; > +const char *dump_file_name; > + > /* Table of tree dump switches. This must be consistent with the > TREE_DUMP_INDEX enumeration in dumpfile.h. */ > static struct dump_file_info dump_files[TDI_end] = > @@ -514,14 +520,6 @@ dump_phase_enabled_p (int phase) > } > } > > -/* Return true if any of the dumps are enabled, false otherwise. */ > - > -inline bool > -dump_enabled_p (void) > -{ > - return (dump_file || alt_dump_file); > -} > - > /* Returns nonzero if tree dump PHASE has been initialized. */ > > int > @@ -836,16 +834,6 @@ opt_info_switch_p (const char *arg) > return opt_info_enable_all ((TDF_TREE | TDF_RTL | TDF_IPA), flags, filename); > } > > -/* Return true if any dumps are enabled for the given MSG_TYPE, false > - otherwise. */ > - > -bool > -dump_kind_p (int msg_type) > -{ > - return (dump_file && (msg_type & pflags)) > - || (alt_dump_file && (msg_type & alt_flags)); > -} > - > /* Print basic block on the dump streams. */ > > void > Index: dumpfile.h > =================================================================== > --- dumpfile.h (revision 192695) > +++ dumpfile.h (working copy) > @@ -108,7 +108,6 @@ struct dump_file_info > int num; /* dump file number */ > }; > > - > /* In dumpfile.c */ > extern char *get_dump_file_name (int); > extern int dump_initialized_p (int); > @@ -120,8 +119,6 @@ extern void dump_node (const_tree, int, FILE *); > extern int dump_switch_p (const char *); > extern int opt_info_switch_p (const char *); > extern const char *dump_flag_name (int); > -extern bool dump_kind_p (int); > -extern inline bool dump_enabled_p (void); > extern void dump_printf (int, const char *, ...) ATTRIBUTE_PRINTF_2; > extern void dump_printf_loc (int, source_location, > const char *, ...) ATTRIBUTE_PRINTF_3; > @@ -142,10 +139,19 @@ extern void dump_bb (FILE *, basic_block, int, int > > /* Global variables used to communicate with passes. */ > extern FILE *dump_file; > +extern FILE *alt_dump_file; > extern int dump_flags; > extern const char *dump_file_name; > > /* Return the dump_file_info for the given phase. */ > extern struct dump_file_info *get_dump_file_info (int); > > +/* Return true if any of the dumps are enabled, false otherwise. */ > + > +static inline bool > +dump_enabled_p (void) > +{ > + return (dump_file || alt_dump_file); > +} > + > #endif /* GCC_DUMPFILE_H */ > Index: toplev.c > =================================================================== > --- toplev.c (revision 192695) > +++ toplev.c (working copy) > @@ -169,8 +169,6 @@ const char *user_label_prefix; > FILE *asm_out_file; > FILE *aux_info_file; > FILE *stack_usage_file = NULL; > -FILE *dump_file = NULL; > -const char *dump_file_name; > > /* The current working directory of a translation. It's generally the > directory from which compilation was initiated, but a preprocessed > Index: tree-vect-loop-manip.c > =================================================================== > --- tree-vect-loop-manip.c (revision 192695) > +++ tree-vect-loop-manip.c (working copy) > @@ -792,7 +792,7 @@ slpeel_make_loop_iterate_ntimes (struct loop *loop > free_stmt_vec_info (orig_cond); > > loop_loc = find_loop_location (loop); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > if (LOCATION_LOCUS (loop_loc) != UNKNOWN_LOC) > dump_printf (MSG_NOTE, "\nloop at %s:%d: ", LOC_FILE (loop_loc), > @@ -1683,7 +1683,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo) > > /* Analyze phi functions of the loop header. */ > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "vect_can_advance_ivs_p:"); > for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi)) > { > @@ -1691,7 +1691,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo) > tree evolution_part; > > phi = gsi_stmt (gsi); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "Analyze phi: "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0); > @@ -1702,7 +1702,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo) > > if (virtual_operand_p (PHI_RESULT (phi))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "virtual phi. skip."); > continue; > @@ -1712,7 +1712,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo) > > if (STMT_VINFO_DEF_TYPE (vinfo_for_stmt (phi)) == vect_reduction_def) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "reduc phi. skip."); > continue; > @@ -1725,13 +1725,13 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo) > > if (!access_fn) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "No Access function."); > return false; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "Access function of PHI: "); > @@ -1742,7 +1742,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo) > > if (evolution_part == NULL_TREE) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf (MSG_MISSED_OPTIMIZATION, "No evolution."); > return false; > } > @@ -1827,7 +1827,7 @@ vect_update_ivs_after_vectorizer (loop_vec_info lo > > phi = gsi_stmt (gsi); > phi1 = gsi_stmt (gsi1); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "vect_update_ivs_after_vectorizer: phi: "); > @@ -1837,7 +1837,7 @@ vect_update_ivs_after_vectorizer (loop_vec_info lo > /* Skip virtual phi's. */ > if (virtual_operand_p (PHI_RESULT (phi))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "virtual phi. skip."); > continue; > @@ -1847,7 +1847,7 @@ vect_update_ivs_after_vectorizer (loop_vec_info lo > stmt_info = vinfo_for_stmt (phi); > if (STMT_VINFO_DEF_TYPE (stmt_info) == vect_reduction_def) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "reduc phi. skip."); > continue; > @@ -1910,7 +1910,7 @@ vect_do_peeling_for_loop_bound (loop_vec_info loop > tree cond_expr = NULL_TREE; > gimple_seq cond_expr_stmt_list = NULL; > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "=== vect_do_peeling_for_loop_bound ==="); > > @@ -2022,7 +2022,7 @@ vect_gen_niters_for_prolog_loop (loop_vec_info loo > { > int npeel = LOOP_PEELING_FOR_ALIGNMENT (loop_vinfo); > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "known peeling = %d.", npeel); > > @@ -2076,7 +2076,7 @@ vect_gen_niters_for_prolog_loop (loop_vec_info loo > if (TREE_CODE (loop_niters) != INTEGER_CST) > iters = fold_build2 (MIN_EXPR, niters_type, iters, loop_niters); > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "niters for prolog loop: "); > @@ -2134,7 +2134,7 @@ vect_update_inits_of_drs (loop_vec_info loop_vinfo > VEC (data_reference_p, heap) *datarefs = LOOP_VINFO_DATAREFS (loop_vinfo); > struct data_reference *dr; > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "=== vect_update_inits_of_dr ==="); > > @@ -2163,7 +2163,7 @@ vect_do_peeling_for_alignment (loop_vec_info loop_ > int max_iter; > int bound = 0; > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "=== vect_do_peeling_for_alignment ==="); > > @@ -2475,7 +2475,7 @@ vect_create_cond_for_alias_checks (loop_vec_info l > segment_length_a = vect_vfa_segment_size (dr_a, length_factor); > segment_length_b = vect_vfa_segment_size (dr_b, length_factor); > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "create runtime check for data references "); > @@ -2506,7 +2506,7 @@ vect_create_cond_for_alias_checks (loop_vec_info l > *cond_expr = part_cond_expr; > } > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "created %u versioning for alias checks.\n", > VEC_length (ddr_p, may_alias_ddrs)); > Index: tree-vectorizer.c > =================================================================== > --- tree-vectorizer.c (revision 192695) > +++ tree-vectorizer.c (working copy) > @@ -107,7 +107,7 @@ vectorize_loops (void) > loop_vec_info loop_vinfo; > vect_location = find_loop_location (loop); > if (LOCATION_LOCUS (vect_location) != UNKNOWN_LOC > - && dump_kind_p (MSG_ALL)) > + && dump_enabled_p ()) > dump_printf (MSG_ALL, "\nAnalyzing loop at %s:%d\n", > LOC_FILE (vect_location), LOC_LINE (vect_location)); > > @@ -118,7 +118,7 @@ vectorize_loops (void) > continue; > > if (LOCATION_LOCUS (vect_location) != UNKNOWN_LOC > - && dump_kind_p (MSG_ALL)) > + && dump_enabled_p ()) > dump_printf (MSG_ALL, "\n\nVectorizing loop at %s:%d\n", > LOC_FILE (vect_location), LOC_LINE (vect_location)); > vect_transform_loop (loop_vinfo); > @@ -128,8 +128,8 @@ vectorize_loops (void) > vect_location = UNKNOWN_LOC; > > statistics_counter_event (cfun, "Vectorized loops", num_vectorized_loops); > - if (dump_kind_p (MSG_ALL) > - || (num_vectorized_loops > 0 && dump_kind_p (MSG_ALL))) > + if (dump_enabled_p () > + || (num_vectorized_loops > 0 && dump_enabled_p ())) > dump_printf_loc (MSG_ALL, vect_location, > "vectorized %u loops in function.\n", > num_vectorized_loops); > @@ -170,7 +170,7 @@ execute_vect_slp (void) > if (vect_slp_analyze_bb (bb)) > { > vect_slp_transform_bb (bb); > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "basic block vectorized using SLP\n"); > } > Index: tree-vect-loop.c > =================================================================== > --- tree-vect-loop.c (revision 192695) > +++ tree-vect-loop.c (working copy) > @@ -187,7 +187,7 @@ vect_determine_vectorization_factor (loop_vec_info > gimple_stmt_iterator pattern_def_si = gsi_none (); > bool analyze_pattern_stmt = false; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_determine_vectorization_factor ==="); > > @@ -199,7 +199,7 @@ vect_determine_vectorization_factor (loop_vec_info > { > phi = gsi_stmt (si); > stmt_info = vinfo_for_stmt (phi); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "==> examining phi: "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0); > @@ -212,7 +212,7 @@ vect_determine_vectorization_factor (loop_vec_info > gcc_assert (!STMT_VINFO_VECTYPE (stmt_info)); > scalar_type = TREE_TYPE (PHI_RESULT (phi)); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "get vectype for scalar type: "); > @@ -222,7 +222,7 @@ vect_determine_vectorization_factor (loop_vec_info > vectype = get_vectype_for_scalar_type (scalar_type); > if (!vectype) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: unsupported " > @@ -234,14 +234,14 @@ vect_determine_vectorization_factor (loop_vec_info > } > STMT_VINFO_VECTYPE (stmt_info) = vectype; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "vectype: "); > dump_generic_expr (MSG_NOTE, TDF_SLIM, vectype); > } > > nunits = TYPE_VECTOR_SUBPARTS (vectype); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "nunits = %d", nunits); > > if (!vectorization_factor > @@ -261,7 +261,7 @@ vect_determine_vectorization_factor (loop_vec_info > > stmt_info = vinfo_for_stmt (stmt); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "==> examining statement: "); > @@ -281,7 +281,7 @@ vect_determine_vectorization_factor (loop_vec_info > { > stmt = pattern_stmt; > stmt_info = vinfo_for_stmt (pattern_stmt); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "==> examining pattern statement: "); > @@ -290,7 +290,7 @@ vect_determine_vectorization_factor (loop_vec_info > } > else > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "skip."); > gsi_next (&si); > continue; > @@ -330,7 +330,7 @@ vect_determine_vectorization_factor (loop_vec_info > > if (!gsi_end_p (pattern_def_si)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "==> examining pattern def stmt: "); > @@ -353,7 +353,7 @@ vect_determine_vectorization_factor (loop_vec_info > > if (gimple_get_lhs (stmt) == NULL_TREE) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: irregular stmt."); > @@ -365,7 +365,7 @@ vect_determine_vectorization_factor (loop_vec_info > > if (VECTOR_MODE_P (TYPE_MODE (gimple_expr_type (stmt)))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: vector stmt in loop:"); > @@ -389,7 +389,7 @@ vect_determine_vectorization_factor (loop_vec_info > { > gcc_assert (!STMT_VINFO_DATA_REF (stmt_info)); > scalar_type = TREE_TYPE (gimple_get_lhs (stmt)); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "get vectype for scalar type: "); > @@ -398,7 +398,7 @@ vect_determine_vectorization_factor (loop_vec_info > vectype = get_vectype_for_scalar_type (scalar_type); > if (!vectype) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: unsupported " > @@ -417,7 +417,7 @@ vect_determine_vectorization_factor (loop_vec_info > support one vector size per loop). */ > scalar_type = vect_get_smallest_scalar_type (stmt, &dummy, > &dummy); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "get vectype for scalar type: "); > @@ -426,7 +426,7 @@ vect_determine_vectorization_factor (loop_vec_info > vf_vectype = get_vectype_for_scalar_type (scalar_type); > if (!vf_vectype) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: unsupported data-type "); > @@ -439,7 +439,7 @@ vect_determine_vectorization_factor (loop_vec_info > if ((GET_MODE_SIZE (TYPE_MODE (vectype)) > != GET_MODE_SIZE (TYPE_MODE (vf_vectype)))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: different sized vector " > @@ -453,14 +453,14 @@ vect_determine_vectorization_factor (loop_vec_info > return false; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "vectype: "); > dump_generic_expr (MSG_NOTE, TDF_SLIM, vf_vectype); > } > > nunits = TYPE_VECTOR_SUBPARTS (vf_vectype); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "nunits = %d", nunits); > if (!vectorization_factor > || (nunits > vectorization_factor)) > @@ -475,12 +475,12 @@ vect_determine_vectorization_factor (loop_vec_info > } > > /* TODO: Analyze cost. Decide if worth while to vectorize. */ > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "vectorization factor = %d", > vectorization_factor); > if (vectorization_factor <= 1) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: unsupported data-type"); > return false; > @@ -517,7 +517,7 @@ vect_is_simple_iv_evolution (unsigned loop_nb, tre > step_expr = evolution_part; > init_expr = unshare_expr (initial_condition_in_loop_num (access_fn, > loop_nb)); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "step: "); > dump_generic_expr (MSG_NOTE, TDF_SLIM, step_expr); > @@ -530,7 +530,7 @@ vect_is_simple_iv_evolution (unsigned loop_nb, tre > > if (TREE_CODE (step_expr) != INTEGER_CST) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "step unknown."); > return false; > @@ -555,7 +555,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v > gimple_stmt_iterator gsi; > bool double_reduc; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_analyze_scalar_cycles ==="); > > @@ -569,7 +569,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v > tree def = PHI_RESULT (phi); > stmt_vec_info stmt_vinfo = vinfo_for_stmt (phi); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "Analyze phi: "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0); > @@ -587,7 +587,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v > if (access_fn) > { > STRIP_NOPS (access_fn); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "Access function of PHI: "); > @@ -606,7 +606,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v > > gcc_assert (STMT_VINFO_LOOP_PHI_EVOLUTION_PART (stmt_vinfo) != > NULL_TREE); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "Detected induction."); > STMT_VINFO_DEF_TYPE (stmt_vinfo) = vect_induction_def; > } > @@ -621,7 +621,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v > gimple reduc_stmt; > bool nested_cycle; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "Analyze phi: "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0); > @@ -637,7 +637,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v > { > if (double_reduc) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Detected double reduction."); > > @@ -649,7 +649,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v > { > if (nested_cycle) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Detected vectorizable nested cycle."); > > @@ -659,7 +659,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v > } > else > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Detected reduction."); > > @@ -675,7 +675,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v > } > } > else > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Unknown def-use cycle pattern."); > } > @@ -737,7 +737,7 @@ vect_get_loop_niters (struct loop *loop, tree *num > { > tree niters; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== get_loop_niters ==="); > niters = number_of_exit_cond_executions (loop); > @@ -747,7 +747,7 @@ vect_get_loop_niters (struct loop *loop, tree *num > { > *number_of_iterations = niters; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "==> get_loop_niters:"); > dump_generic_expr (MSG_NOTE, TDF_SLIM, *number_of_iterations); > @@ -995,7 +995,7 @@ vect_analyze_loop_1 (struct loop *loop) > { > loop_vec_info loop_vinfo; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "===== analyze_loop_nest_1 ====="); > > @@ -1004,7 +1004,7 @@ vect_analyze_loop_1 (struct loop *loop) > loop_vinfo = vect_analyze_loop_form (loop); > if (!loop_vinfo) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "bad inner-loop form."); > return NULL; > @@ -1030,7 +1030,7 @@ vect_analyze_loop_form (struct loop *loop) > tree number_of_iterations = NULL; > loop_vec_info inner_loop_vinfo = NULL; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_analyze_loop_form ==="); > > @@ -1054,7 +1054,7 @@ vect_analyze_loop_form (struct loop *loop) > > if (loop->num_nodes != 2) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: control flow in loop."); > return NULL; > @@ -1062,7 +1062,7 @@ vect_analyze_loop_form (struct loop *loop) > > if (empty_block_p (loop->header)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: empty loop."); > return NULL; > @@ -1092,7 +1092,7 @@ vect_analyze_loop_form (struct loop *loop) > > if ((loop->inner)->inner || (loop->inner)->next) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: multiple nested loops."); > return NULL; > @@ -1102,7 +1102,7 @@ vect_analyze_loop_form (struct loop *loop) > inner_loop_vinfo = vect_analyze_loop_1 (loop->inner); > if (!inner_loop_vinfo) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: Bad inner loop."); > return NULL; > @@ -1111,7 +1111,7 @@ vect_analyze_loop_form (struct loop *loop) > if (!expr_invariant_in_loop_p (loop, > LOOP_VINFO_NITERS (inner_loop_vinfo))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: inner-loop count not invariant."); > destroy_loop_vec_info (inner_loop_vinfo, true); > @@ -1120,7 +1120,7 @@ vect_analyze_loop_form (struct loop *loop) > > if (loop->num_nodes != 5) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: control flow in loop."); > destroy_loop_vec_info (inner_loop_vinfo, true); > @@ -1136,14 +1136,14 @@ vect_analyze_loop_form (struct loop *loop) > || !single_exit (innerloop) > || single_exit (innerloop)->dest != EDGE_PRED (loop->latch, 0)->src) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: unsupported outerloop form."); > destroy_loop_vec_info (inner_loop_vinfo, true); > return NULL; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Considering outer-loop vectorization."); > } > @@ -1151,7 +1151,7 @@ vect_analyze_loop_form (struct loop *loop) > if (!single_exit (loop) > || EDGE_COUNT (loop->header->preds) != 2) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > if (!single_exit (loop)) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > @@ -1172,7 +1172,7 @@ vect_analyze_loop_form (struct loop *loop) > if (!empty_block_p (loop->latch) > || !gimple_seq_empty_p (phi_nodes (loop->latch))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: unexpected loop form."); > if (inner_loop_vinfo) > @@ -1187,12 +1187,12 @@ vect_analyze_loop_form (struct loop *loop) > if (!(e->flags & EDGE_ABNORMAL)) > { > split_loop_exit_edge (e); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf (MSG_NOTE, "split exit edge."); > } > else > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: abnormal loop exit edge."); > if (inner_loop_vinfo) > @@ -1204,7 +1204,7 @@ vect_analyze_loop_form (struct loop *loop) > loop_cond = vect_get_loop_niters (loop, &number_of_iterations); > if (!loop_cond) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: complicated exit condition."); > if (inner_loop_vinfo) > @@ -1214,7 +1214,7 @@ vect_analyze_loop_form (struct loop *loop) > > if (!number_of_iterations) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: number of iterations cannot be " > "computed."); > @@ -1225,7 +1225,7 @@ vect_analyze_loop_form (struct loop *loop) > > if (chrec_contains_undetermined (number_of_iterations)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Infinite number of iterations."); > if (inner_loop_vinfo) > @@ -1235,7 +1235,7 @@ vect_analyze_loop_form (struct loop *loop) > > if (!NITERS_KNOWN_P (number_of_iterations)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "Symbolic number of iterations is "); > @@ -1244,7 +1244,7 @@ vect_analyze_loop_form (struct loop *loop) > } > else if (TREE_INT_CST_LOW (number_of_iterations) == 0) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: number of iterations = 0."); > if (inner_loop_vinfo) > @@ -1292,7 +1292,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v > HOST_WIDE_INT estimated_niter; > int min_profitable_estimate; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_analyze_loop_operations ==="); > > @@ -1328,7 +1328,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v > LOOP_VINFO_SLP_UNROLLING_FACTOR (loop_vinfo)); > > LOOP_VINFO_VECT_FACTOR (loop_vinfo) = vectorization_factor; > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Updating vectorization factor to %d ", > vectorization_factor); > @@ -1344,7 +1344,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v > ok = true; > > stmt_info = vinfo_for_stmt (phi); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "examining phi: "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0); > @@ -1363,7 +1363,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v > && STMT_VINFO_DEF_TYPE (stmt_info) > != vect_double_reduction_def) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Unsupported loop-closed phi in " > "outer-loop."); > @@ -1405,7 +1405,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v > if (STMT_VINFO_LIVE_P (stmt_info)) > { > /* FORNOW: not yet supported. */ > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: value used after loop."); > return false; > @@ -1415,7 +1415,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v > && STMT_VINFO_DEF_TYPE (stmt_info) != vect_induction_def) > { > /* A scalar-dependence cycle that we don't support. */ > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: scalar dependence cycle."); > return false; > @@ -1430,7 +1430,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v > > if (!ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: relevant phi not " > @@ -1456,18 +1456,17 @@ vect_analyze_loop_operations (loop_vec_info loop_v > touching this loop. */ > if (!need_to_vectorize) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "All the computation can be taken out of the loop."); > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: redundant loop. no profit to " > "vectorize."); > return false; > } > > - if (LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo) > - && dump_kind_p (MSG_NOTE)) > + if (LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo) && dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vectorization_factor = %d, niters = " > HOST_WIDE_INT_PRINT_DEC, vectorization_factor, > @@ -1478,10 +1477,10 @@ vect_analyze_loop_operations (loop_vec_info loop_v > || ((max_niter = max_stmt_executions_int (loop)) != -1 > && (unsigned HOST_WIDE_INT) max_niter < vectorization_factor)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: iteration count too small."); > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: iteration count smaller than " > "vectorization factor."); > @@ -1500,10 +1499,10 @@ vect_analyze_loop_operations (loop_vec_info loop_v > > if (min_profitable_iters < 0) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: vectorization not profitable."); > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: vector version will never be " > "profitable."); > @@ -1526,10 +1525,10 @@ vect_analyze_loop_operations (loop_vec_info loop_v > if (LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo) > && LOOP_VINFO_INT_NITERS (loop_vinfo) <= th) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: vectorization not profitable."); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "not vectorized: iteration count smaller than user " > "specified loop bound parameter or minimum profitable " > @@ -1541,11 +1540,11 @@ vect_analyze_loop_operations (loop_vec_info loop_v > && ((unsigned HOST_WIDE_INT) estimated_niter > <= MAX (th, (unsigned)min_profitable_estimate))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: estimated iteration count too " > "small."); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "not vectorized: estimated iteration count smaller " > "than specified loop bound parameter or minimum " > @@ -1558,18 +1557,18 @@ vect_analyze_loop_operations (loop_vec_info loop_v > || LOOP_VINFO_INT_NITERS (loop_vinfo) % vectorization_factor != 0 > || LOOP_PEELING_FOR_ALIGNMENT (loop_vinfo)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "epilog loop required."); > if (!vect_can_advance_ivs_p (loop_vinfo)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: can't create epilog loop 1."); > return false; > } > if (!slpeel_can_duplicate_loop_p (loop, single_exit (loop))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: can't create epilog loop 2."); > return false; > @@ -1602,7 +1601,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) > ok = vect_analyze_data_refs (loop_vinfo, NULL, &min_vf); > if (!ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "bad data references."); > return false; > @@ -1620,7 +1619,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) > ok = vect_mark_stmts_to_be_vectorized (loop_vinfo); > if (!ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "unexpected pattern."); > return false; > @@ -1635,7 +1634,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) > if (!ok > || max_vf < min_vf) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "bad data dependence."); > return false; > @@ -1644,14 +1643,14 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) > ok = vect_determine_vectorization_factor (loop_vinfo); > if (!ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "can't determine vectorization factor."); > return false; > } > if (max_vf < LOOP_VINFO_VECT_FACTOR (loop_vinfo)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "bad data dependence."); > return false; > @@ -1663,7 +1662,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) > ok = vect_analyze_data_refs_alignment (loop_vinfo, NULL); > if (!ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "bad data alignment."); > return false; > @@ -1675,7 +1674,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) > ok = vect_analyze_data_ref_accesses (loop_vinfo, NULL); > if (!ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "bad data access."); > return false; > @@ -1687,7 +1686,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) > ok = vect_prune_runtime_alias_test_list (loop_vinfo); > if (!ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "too long list of versioning for alias " > "run-time tests."); > @@ -1700,7 +1699,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) > ok = vect_enhance_data_refs_alignment (loop_vinfo); > if (!ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "bad data alignment."); > return false; > @@ -1725,7 +1724,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) > ok = vect_analyze_loop_operations (loop_vinfo, slp); > if (!ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "bad operation or unsupported loop bound."); > return false; > @@ -1749,7 +1748,7 @@ vect_analyze_loop (struct loop *loop) > current_vector_size = 0; > vector_sizes = targetm.vectorize.autovectorize_vector_sizes (); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "===== analyze_loop_nest ====="); > > @@ -1757,7 +1756,7 @@ vect_analyze_loop (struct loop *loop) > && loop_vec_info_for_loop (loop_outer (loop)) > && LOOP_VINFO_VECTORIZABLE_P (loop_vec_info_for_loop > (loop_outer (loop)))) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "outer-loop already vectorized."); > return NULL; > @@ -1769,7 +1768,7 @@ vect_analyze_loop (struct loop *loop) > loop_vinfo = vect_analyze_loop_form (loop); > if (!loop_vinfo) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "bad loop form."); > return NULL; > @@ -1791,7 +1790,7 @@ vect_analyze_loop (struct loop *loop) > > /* Try the next biggest vector size. */ > current_vector_size = 1 << floor_log2 (vector_sizes); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "***** Re-trying analysis with " > "vector size %d\n", current_vector_size); > @@ -2023,7 +2022,7 @@ vect_is_slp_reduction (loop_vec_info loop_info, gi > == vect_internal_def > && !is_loop_header_bb_p (gimple_bb (def_stmt))))) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "swapping oprnds: "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, next_stmt, 0); > @@ -2125,7 +2124,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > > if (!flow_bb_inside_loop_p (loop, gimple_bb (use_stmt))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "intermediate value used outside loop."); > > @@ -2137,7 +2136,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > nloop_uses++; > if (nloop_uses > 1) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "reduction used in loop."); > return NULL; > @@ -2146,7 +2145,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > > if (TREE_CODE (loop_arg) != SSA_NAME) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "reduction: not ssa_name: "); > @@ -2158,7 +2157,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > def_stmt = SSA_NAME_DEF_STMT (loop_arg); > if (!def_stmt) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "reduction: no def_stmt."); > return NULL; > @@ -2166,7 +2165,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > > if (!is_gimple_assign (def_stmt) && gimple_code (def_stmt) != GIMPLE_PHI) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, def_stmt, 0); > return NULL; > } > @@ -2194,7 +2193,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > nloop_uses++; > if (nloop_uses > 1) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "reduction used in loop."); > return NULL; > @@ -2210,7 +2209,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > if (gimple_phi_num_args (def_stmt) != 1 > || TREE_CODE (op1) != SSA_NAME) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "unsupported phi node definition."); > > @@ -2223,7 +2222,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > && flow_bb_inside_loop_p (loop->inner, gimple_bb (def1)) > && is_gimple_assign (def1)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > report_vect_op (MSG_NOTE, def_stmt, > "detected double reduction: "); > > @@ -2250,7 +2249,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > if (check_reduction > && (!commutative_tree_code (code) || !associative_tree_code (code))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, > "reduction: not commutative/associative: "); > return NULL; > @@ -2260,7 +2259,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > { > if (code != COND_EXPR) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, > "reduction: not binary operation: "); > > @@ -2279,7 +2278,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > > if (TREE_CODE (op1) != SSA_NAME && TREE_CODE (op2) != SSA_NAME) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, > "reduction: uses not ssa_names: "); > > @@ -2293,7 +2292,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > > if (TREE_CODE (op1) != SSA_NAME && TREE_CODE (op2) != SSA_NAME) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, > "reduction: uses not ssa_names: "); > > @@ -2311,7 +2310,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > || (op4 && TREE_CODE (op4) == SSA_NAME > && !types_compatible_p (type, TREE_TYPE (op4)))) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "reduction: multiple types: operation type: "); > @@ -2353,7 +2352,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > && check_reduction) > { > /* Changing the order of operations changes the semantics. */ > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, > "reduction: unsafe fp math optimization: "); > return NULL; > @@ -2362,7 +2361,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > && check_reduction) > { > /* Changing the order of operations changes the semantics. */ > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, > "reduction: unsafe int math optimization: "); > return NULL; > @@ -2370,7 +2369,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > else if (SAT_FIXED_POINT_TYPE_P (type) && check_reduction) > { > /* Changing the order of operations changes the semantics. */ > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, > "reduction: unsafe fixed-point math optimization: "); > return NULL; > @@ -2407,7 +2406,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > if (code != COND_EXPR > && ((!def1 || gimple_nop_p (def1)) && (!def2 || gimple_nop_p (def2)))) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > report_vect_op (MSG_NOTE, def_stmt, "reduction: no defs for operands: "); > return NULL; > } > @@ -2429,7 +2428,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > == vect_internal_def > && !is_loop_header_bb_p (gimple_bb (def1))))))) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > report_vect_op (MSG_NOTE, def_stmt, "detected reduction: "); > return def_stmt; > } > @@ -2452,7 +2451,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > /* Swap operands (just for simplicity - so that the rest of the code > can assume that the reduction variable is always the last (second) > argument). */ > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > report_vect_op (MSG_NOTE, def_stmt, > "detected reduction: need to swap operands: "); > > @@ -2464,7 +2463,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > } > else > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > report_vect_op (MSG_NOTE, def_stmt, "detected reduction: "); > } > > @@ -2474,14 +2473,14 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf > /* Try to find SLP reduction chain. */ > if (check_reduction && vect_is_slp_reduction (loop_info, phi, def_stmt)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > report_vect_op (MSG_NOTE, def_stmt, > "reduction: detected reduction chain: "); > > return def_stmt; > } > > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, > "reduction: unknown pattern: "); > > @@ -2589,7 +2588,7 @@ vect_get_known_peeling_cost (loop_vec_info loop_vi > if (!LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo)) > { > *peel_iters_epilogue = vf/2; > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "cost model: epilogue peel iters set to vf/2 " > "because loop iterations are unknown ."); > @@ -2882,7 +2881,7 @@ vect_estimate_min_profitable_iters (loop_vec_info > /* vector version will never be profitable. */ > else > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "cost model: the vector iteration cost = %d " > "divided by the scalar iteration cost = %d " > @@ -2893,7 +2892,7 @@ vect_estimate_min_profitable_iters (loop_vec_info > return; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "Cost model analysis: \n"); > dump_printf (MSG_NOTE, " Vector inside of loop cost: %d\n", > @@ -2925,7 +2924,7 @@ vect_estimate_min_profitable_iters (loop_vec_info > then skip the vectorized loop. */ > min_profitable_iters--; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > " Runtime profitability threshold = %d\n", > min_profitable_iters); > > @@ -2950,7 +2949,7 @@ vect_estimate_min_profitable_iters (loop_vec_info > } > min_profitable_estimate --; > min_profitable_estimate = MAX (min_profitable_estimate, > min_profitable_iters); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > " Static estimate profitability threshold = %d\n", > min_profitable_iters); > @@ -3010,7 +3009,7 @@ vect_model_reduction_cost (stmt_vec_info stmt_info > vectype = get_vectype_for_scalar_type (TREE_TYPE (reduction_op)); > if (!vectype) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "unsupported data-type "); > @@ -3081,7 +3080,7 @@ vect_model_reduction_cost (stmt_vec_info stmt_info > } > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf (MSG_NOTE, > "vect_model_reduction_cost: inside_cost = %d, " > "prologue_cost = %d, epilogue_cost = %d .", inside_cost, > @@ -3110,7 +3109,7 @@ vect_model_induction_cost (stmt_vec_info stmt_info > prologue_cost = add_stmt_cost (target_cost_data, 2, scalar_to_vec, > stmt_info, 0, vect_prologue); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_induction_cost: inside_cost = %d, " > "prologue_cost = %d .", inside_cost, prologue_cost); > @@ -3239,7 +3238,7 @@ get_initial_def_for_induction (gimple iv_phi) > new_bb = gsi_insert_on_edge_immediate (pe, init_stmt); > gcc_assert (!new_bb); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "created new init_stmt: "); > @@ -3382,7 +3381,7 @@ get_initial_def_for_induction (gimple iv_phi) > && !STMT_VINFO_LIVE_P (stmt_vinfo)); > > STMT_VINFO_VEC_STMT (stmt_vinfo) = new_stmt; > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "vector of inductions after inner-loop:"); > @@ -3392,7 +3391,7 @@ get_initial_def_for_induction (gimple iv_phi) > } > > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "transform induction: created def-use cycle: "); > @@ -3800,7 +3799,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap) > > add_phi_arg (phi, def, loop_latch_edge (loop), UNKNOWN_LOCATION); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "transform reduction: created def-use cycle: "); > @@ -4001,7 +4000,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap) > /*** Case 1: Create: > v_out2 = reduc_expr <v_out1> */ > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Reduce using direct vector reduction."); > > @@ -4052,7 +4051,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap) > Create: va = vop <va, va'> > } */ > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Reduce using vector shifts"); > > @@ -4093,7 +4092,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap) > Create: s = op <s, s'> // For non SLP cases > } */ > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Reduce using scalar code. "); > > @@ -4184,7 +4183,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap) > { > tree rhs; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "extract scalar result"); > > @@ -4423,7 +4422,7 @@ vect_finalize_reduction: > UNKNOWN_LOCATION); > add_phi_arg (vect_phi, PHI_RESULT (inner_phi), > loop_latch_edge (outer_loop), UNKNOWN_LOCATION); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "created double reduction phi node: "); > @@ -4773,7 +4772,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i > { > if (!vectorizable_condition (stmt, gsi, NULL, ops[reduc_index], 0, NULL)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "unsupported condition in reduction"); > > @@ -4788,7 +4787,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i > optab = optab_for_tree_code (code, vectype_in, optab_default); > if (!optab) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "no optab."); > > @@ -4797,7 +4796,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i > > if (optab_handler (optab, vec_mode) == CODE_FOR_nothing) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf (MSG_NOTE, "op not supported by target."); > > if (GET_MODE_SIZE (vec_mode) != UNITS_PER_WORD > @@ -4805,7 +4804,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i > < vect_min_worthwhile_factor (code)) > return false; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf (MSG_NOTE, "proceeding using word mode."); > } > > @@ -4814,7 +4813,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i > && LOOP_VINFO_VECT_FACTOR (loop_vinfo) > < vect_min_worthwhile_factor (code)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not worthwhile without SIMD support."); > > @@ -4895,7 +4894,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i > optab_default); > if (!reduc_optab) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "no optab for reduction."); > > @@ -4905,7 +4904,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i > if (reduc_optab > && optab_handler (reduc_optab, vec_mode) == CODE_FOR_nothing) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "reduc op not supported by target."); > > @@ -4916,7 +4915,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i > { > if (!nested_cycle || double_reduc) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "no reduc code for scalar code."); > > @@ -4926,7 +4925,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i > > if (double_reduc && ncopies > 1) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "multiple types in double reduction"); > > @@ -4945,7 +4944,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i > ops[1] = fold_convert (TREE_TYPE (ops[0]), ops[1]); > else > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "invalid types in dot-prod"); > > @@ -4963,7 +4962,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i > > /** Transform. **/ > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "transform reduction."); > > /* FORNOW: Multiple types are not supported for condition. */ > @@ -5249,7 +5248,7 @@ vectorizable_induction (gimple phi, gimple_stmt_it > > if (ncopies > 1) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "multiple types in nested loop."); > return false; > @@ -5273,7 +5272,7 @@ vectorizable_induction (gimple phi, gimple_stmt_it > if (!(STMT_VINFO_RELEVANT_P (exit_phi_vinfo) > && !STMT_VINFO_LIVE_P (exit_phi_vinfo))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "inner-loop induction only used outside " > "of the outer vectorized loop."); > @@ -5297,7 +5296,7 @@ vectorizable_induction (gimple phi, gimple_stmt_it > if (!vec_stmt) /* transformation not required. */ > { > STMT_VINFO_TYPE (stmt_info) = induc_vec_info_type; > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vectorizable_induction ==="); > vect_model_induction_cost (stmt_info, ncopies); > @@ -5306,7 +5305,7 @@ vectorizable_induction (gimple phi, gimple_stmt_it > > /** Transform. **/ > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "transform induction phi."); > > vec_def = get_initial_def_for_induction (phi); > @@ -5371,7 +5370,7 @@ vectorizable_live_operation (gimple stmt, > && !vect_is_simple_use (op, stmt, loop_vinfo, NULL, &def_stmt, &def, > &dt)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "use not simple."); > return false; > @@ -5410,7 +5409,7 @@ vect_loop_kill_debug_uses (struct loop *loop, gimp > { > if (gimple_debug_bind_p (ustmt)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "killing debug use"); > > @@ -5450,7 +5449,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) > bool check_profitability = false; > int th; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "=== vec_transform_loop ==="); > > /* Use the more conservative vectorization threshold. If the number > @@ -5464,7 +5463,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) > if (th >= LOOP_VINFO_VECT_FACTOR (loop_vinfo) - 1 > && !LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Profitability threshold is %d loop iterations.", th); > check_profitability = true; > @@ -5525,7 +5524,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) > for (si = gsi_start_phis (bb); !gsi_end_p (si); gsi_next (&si)) > { > phi = gsi_stmt (si); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "------>vectorizing phi: "); > @@ -5544,12 +5543,12 @@ vect_transform_loop (loop_vec_info loop_vinfo) > > if ((TYPE_VECTOR_SUBPARTS (STMT_VINFO_VECTYPE (stmt_info)) > != (unsigned HOST_WIDE_INT) vectorization_factor) > - && dump_kind_p (MSG_NOTE)) > + && dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "multiple-types."); > > if (STMT_VINFO_DEF_TYPE (stmt_info) == vect_induction_def) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "transform phi."); > vect_transform_stmt (phi, NULL, NULL, NULL, NULL); > } > @@ -5565,7 +5564,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) > else > stmt = gsi_stmt (si); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "------>vectorizing statement: "); > @@ -5637,7 +5636,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) > > if (!gsi_end_p (pattern_def_si)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "==> vectorizing pattern def " > @@ -5664,7 +5663,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) > STMT_VINFO_VECTYPE (stmt_info)); > if (!STMT_SLP_TYPE (stmt_info) > && nunits != (unsigned int) vectorization_factor > - && dump_kind_p (MSG_NOTE)) > + && dump_enabled_p ()) > /* For SLP VF is set according to unrolling factor, and not to > vector size, hence for SLP this print is not valid. */ > dump_printf_loc (MSG_NOTE, vect_location, > @@ -5678,7 +5677,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) > { > slp_scheduled = true; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== scheduling SLP instances ==="); > > @@ -5698,7 +5697,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) > } > > /* -------- vectorize statement ------------ */ > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "transform statement."); > > grouped_store = false; > @@ -5741,9 +5740,9 @@ vect_transform_loop (loop_vec_info loop_vinfo) > until all the loops have been transformed? */ > update_ssa (TODO_update_ssa); > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "LOOP > VECTORIZED."); > - if (loop->inner && dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (loop->inner && dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "OUTER LOOP VECTORIZED."); > } > Index: tree-vect-data-refs.c > =================================================================== > --- tree-vect-data-refs.c (revision 192695) > +++ tree-vect-data-refs.c (working copy) > @@ -60,7 +60,7 @@ vect_lanes_optab_supported_p (const char *name, co > > if (array_mode == BLKmode) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "no array mode for %s[" HOST_WIDE_INT_PRINT_DEC "]", > GET_MODE_NAME (mode), count); > @@ -69,14 +69,14 @@ vect_lanes_optab_supported_p (const char *name, co > > if (convert_optab_handler (optab, array_mode, mode) == CODE_FOR_nothing) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "cannot use %s<%s><%s>", name, > GET_MODE_NAME (array_mode), GET_MODE_NAME (mode)); > return false; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "can use %s<%s><%s>", name, GET_MODE_NAME (array_mode), > GET_MODE_NAME (mode)); > @@ -439,7 +439,7 @@ vect_check_interleaving (struct data_reference *dr > if (diff_mod_size == 0) > { > vect_update_interleaving_chain (drb, dra); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "Detected interleaving "); > @@ -462,7 +462,7 @@ vect_check_interleaving (struct data_reference *dr > if (diff_mod_size == 0) > { > vect_update_interleaving_chain (dra, drb); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "Detected interleaving "); > @@ -524,7 +524,7 @@ vect_mark_for_runtime_alias_test (ddr_p ddr, loop_ > if ((unsigned) PARAM_VALUE (PARAM_VECT_MAX_VERSION_FOR_ALIAS_CHECKS) == 0) > return false; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "mark for run-time aliasing test between "); > @@ -535,7 +535,7 @@ vect_mark_for_runtime_alias_test (ddr_p ddr, loop_ > > if (optimize_loop_nest_for_size_p (loop)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "versioning not supported when optimizing for size."); > return false; > @@ -544,7 +544,7 @@ vect_mark_for_runtime_alias_test (ddr_p ddr, loop_ > /* FORNOW: We don't support versioning with outer-loop vectorization. */ > if (loop->inner) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "versioning not yet supported for outer-loops."); > return false; > @@ -555,7 +555,7 @@ vect_mark_for_runtime_alias_test (ddr_p ddr, loop_ > if (TREE_CODE (DR_STEP (DDR_A (ddr))) != INTEGER_CST > || TREE_CODE (DR_STEP (DDR_B (ddr))) != INTEGER_CST) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "versioning not yet supported for non-constant " > "step"); > @@ -611,7 +611,7 @@ vect_analyze_data_ref_dependence (struct data_depe > > if (loop_vinfo) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "versioning for alias required: " > @@ -637,7 +637,7 @@ vect_analyze_data_ref_dependence (struct data_depe > if (DR_IS_READ (dra) && DR_IS_READ (drb)) > return false; > > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "can't determine dependence between "); > @@ -666,7 +666,7 @@ vect_analyze_data_ref_dependence (struct data_depe > if (dra != drb && vect_check_interleaving (dra, drb)) > return false; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "determined dependence between "); > @@ -686,7 +686,7 @@ vect_analyze_data_ref_dependence (struct data_depe > /* Loop-based vectorization and known data dependence. */ > if (DDR_NUM_DIST_VECTS (ddr) == 0) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "versioning for alias required: " > @@ -704,13 +704,13 @@ vect_analyze_data_ref_dependence (struct data_depe > { > int dist = dist_v[loop_depth]; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "dependence distance = %d.", dist); > > if (dist == 0) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "dependence distance == 0 between "); > @@ -737,7 +737,7 @@ vect_analyze_data_ref_dependence (struct data_depe > /* If DDR_REVERSED_P the order of the data-refs in DDR was > reversed (to make distance vector positive), and the actual > distance is negative. */ > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "dependence distance negative."); > continue; > @@ -749,7 +749,7 @@ vect_analyze_data_ref_dependence (struct data_depe > /* The dependence distance requires reduction of the maximal > vectorization factor. */ > *max_vf = abs (dist); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "adjusting maximal vectorization factor to %i", > *max_vf); > @@ -759,13 +759,13 @@ vect_analyze_data_ref_dependence (struct data_depe > { > /* Dependence distance does not create dependence, as far as > vectorization is concerned, in this case. */ > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "dependence distance >= VF."); > continue; > } > > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized, possible dependence " > @@ -795,7 +795,7 @@ vect_analyze_data_ref_dependences (loop_vec_info l > VEC (ddr_p, heap) *ddrs = NULL; > struct data_dependence_relation *ddr; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_analyze_dependences ==="); > if (loop_vinfo) > @@ -837,7 +837,7 @@ vect_compute_data_ref_alignment (struct data_refer > tree misalign; > tree aligned_to, alignment; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_compute_data_ref_alignment:"); > > @@ -870,7 +870,7 @@ vect_compute_data_ref_alignment (struct data_refer > > if (dr_step % GET_MODE_SIZE (TYPE_MODE (vectype)) == 0) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "inner step divides the vector-size."); > misalign = STMT_VINFO_DR_INIT (stmt_info); > @@ -879,7 +879,7 @@ vect_compute_data_ref_alignment (struct data_refer > } > else > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "inner step doesn't divide the vector-size."); > misalign = NULL_TREE; > @@ -898,7 +898,7 @@ vect_compute_data_ref_alignment (struct data_refer > > if (dr_step % GET_MODE_SIZE (TYPE_MODE (vectype)) != 0) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "SLP: step doesn't divide the vector-size."); > misalign = NULL_TREE; > @@ -911,7 +911,7 @@ vect_compute_data_ref_alignment (struct data_refer > if ((aligned_to && tree_int_cst_compare (aligned_to, alignment) < 0) > || !misalign) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Unknown alignment for access: "); > @@ -941,7 +941,7 @@ vect_compute_data_ref_alignment (struct data_refer > if (!vect_can_force_dr_alignment_p (base, TYPE_ALIGN (vectype)) > || (TREE_STATIC (base) && flag_section_anchors)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "can't force alignment of ref: "); > @@ -953,7 +953,7 @@ vect_compute_data_ref_alignment (struct data_refer > /* Force the alignment of the decl. > NOTE: This is the only change to the code we make during > the analysis phase, before deciding to vectorize the loop. */ > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "force alignment of "); > dump_generic_expr (MSG_NOTE, TDF_SLIM, ref); > @@ -987,7 +987,7 @@ vect_compute_data_ref_alignment (struct data_refer > if (!host_integerp (misalign, 1)) > { > /* Negative or overflowed misalignment value. */ > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "unexpected misalign value"); > return false; > @@ -995,7 +995,7 @@ vect_compute_data_ref_alignment (struct data_refer > > SET_DR_MISALIGNMENT (dr, TREE_INT_CST_LOW (misalign)); > > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "misalign = %d bytes of ref ", DR_MISALIGNMENT (dr)); > @@ -1095,7 +1095,7 @@ vect_update_misalignment_for_peel (struct data_ref > return; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "Setting misalignment to -1."); > SET_DR_MISALIGNMENT (dr, -1); > } > @@ -1142,7 +1142,7 @@ vect_verify_datarefs_alignment (loop_vec_info loop > supportable_dr_alignment = vect_supportable_dr_alignment (dr, false); > if (!supportable_dr_alignment) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > if (DR_IS_READ (dr)) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > @@ -1157,8 +1157,7 @@ vect_verify_datarefs_alignment (loop_vec_info loop > } > return false; > } > - if (supportable_dr_alignment != dr_aligned > - && dump_kind_p (MSG_NOTE)) > + if (supportable_dr_alignment != dr_aligned && dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Vectorizing an unaligned access."); > } > @@ -1215,7 +1214,7 @@ vector_alignment_reachable_p (struct data_referenc > { > HOST_WIDE_INT elmsize = > int_cst_value (TYPE_SIZE_UNIT (TREE_TYPE (vectype))); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "data size =" HOST_WIDE_INT_PRINT_DEC, elmsize); > @@ -1224,7 +1223,7 @@ vector_alignment_reachable_p (struct data_referenc > } > if (DR_MISALIGNMENT (dr) % elmsize) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "data size does not divide the misalignment.\n"); > return false; > @@ -1235,7 +1234,7 @@ vector_alignment_reachable_p (struct data_referenc > { > tree type = TREE_TYPE (DR_REF (dr)); > bool is_packed = not_size_aligned (DR_REF (dr)); > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Unknown misalignment, is_packed = %d",is_packed); > if (targetm.vectorize.vector_alignment_reachable (type, is_packed)) > @@ -1269,7 +1268,7 @@ vect_get_data_access_cost (struct data_reference * > else > vect_get_store_cost (dr, ncopies, inside_cost, body_cost_vec); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_get_data_access_cost: inside_cost = %d, " > "outside_cost = %d.", *inside_cost, *outside_cost); > @@ -1567,7 +1566,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo > unsigned int nelements, mis, same_align_drs_max = 0; > stmt_vector_for_cost body_cost_vec = NULL; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_enhance_data_refs_alignment ==="); > > @@ -1622,7 +1621,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo > and so we can't generate the new base for the pointer. */ > if (STMT_VINFO_STRIDE_LOAD_P (stmt_info)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "strided load prevents peeling"); > do_peeling = false; > @@ -1738,7 +1737,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo > { > if (!aligned_access_p (dr)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "vector alignment may not be reachable"); > break; > @@ -1879,7 +1878,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo > if (STMT_VINFO_GROUPED_ACCESS (stmt_info)) > npeel /= GROUP_SIZE (stmt_info); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Try peeling by %d", npeel); > } > @@ -1951,7 +1950,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo > else > LOOP_PEELING_FOR_ALIGNMENT (loop_vinfo) = DR_MISALIGNMENT (dr0); > SET_DR_MISALIGNMENT (dr0, 0); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "Alignment of access forced using peeling."); > @@ -2077,12 +2076,12 @@ vect_enhance_data_refs_alignment (loop_vec_info lo > stmt_vec_info stmt_info = vinfo_for_stmt (stmt); > dr = STMT_VINFO_DATA_REF (stmt_info); > SET_DR_MISALIGNMENT (dr, 0); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Alignment of access forced using versioning."); > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Versioning for alignment will be applied."); > > @@ -2148,7 +2147,7 @@ vect_find_same_alignment_drs (struct data_dependen > { > int dist = dist_v[loop_depth]; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "dependence distance = %d.", dist); > > @@ -2159,7 +2158,7 @@ vect_find_same_alignment_drs (struct data_dependen > /* Two references with distance zero have the same alignment. */ > VEC_safe_push (dr_p, heap, STMT_VINFO_SAME_ALIGN_REFS (stmtinfo_a), drb); > VEC_safe_push (dr_p, heap, STMT_VINFO_SAME_ALIGN_REFS (stmtinfo_b), dra); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "accesses have the same alignment."); > @@ -2183,7 +2182,7 @@ bool > vect_analyze_data_refs_alignment (loop_vec_info loop_vinfo, > bb_vec_info bb_vinfo) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_analyze_data_refs_alignment ==="); > > @@ -2201,7 +2200,7 @@ vect_analyze_data_refs_alignment (loop_vec_info lo > > if (!vect_compute_data_refs_alignment (loop_vinfo, bb_vinfo)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: can't calculate alignment " > "for data ref."); > @@ -2254,7 +2253,7 @@ vect_analyze_group_access (struct data_reference * > { > GROUP_FIRST_ELEMENT (vinfo_for_stmt (stmt)) = stmt; > GROUP_SIZE (vinfo_for_stmt (stmt)) = groupsize; > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "Detected single element interleaving "); > @@ -2265,13 +2264,13 @@ vect_analyze_group_access (struct data_reference * > > if (loop_vinfo) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Data access with gaps requires scalar " > "epilogue loop"); > if (loop->inner) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Peeling for outer loop is not" > " supported"); > @@ -2284,7 +2283,7 @@ vect_analyze_group_access (struct data_reference * > return true; > } > > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not consecutive access "); > @@ -2324,7 +2323,7 @@ vect_analyze_group_access (struct data_reference * > { > if (DR_IS_WRITE (data_ref)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Two store stmts share the same dr."); > return false; > @@ -2335,7 +2334,7 @@ vect_analyze_group_access (struct data_reference * > if (GROUP_READ_WRITE_DEPENDENCE (vinfo_for_stmt (next)) > || GROUP_READ_WRITE_DEPENDENCE (vinfo_for_stmt (prev))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "READ_WRITE dependence in interleaving."); > return false; > @@ -2355,7 +2354,7 @@ vect_analyze_group_access (struct data_reference * > next_step = DR_STEP (STMT_VINFO_DATA_REF (vinfo_for_stmt (next))); > if (tree_int_cst_compare (step, next_step)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not consecutive access in interleaving"); > return false; > @@ -2372,7 +2371,7 @@ vect_analyze_group_access (struct data_reference * > slp_impossible = true; > if (DR_IS_WRITE (data_ref)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "interleaved store with gaps"); > return false; > @@ -2401,7 +2400,7 @@ vect_analyze_group_access (struct data_reference * > greater than STEP. */ > if (dr_step && dr_step < count_in_bytes + gaps * type_size) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "interleaving size is greater than step for "); > @@ -2424,7 +2423,7 @@ vect_analyze_group_access (struct data_reference * > } > else > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "interleaved store with gaps"); > return false; > @@ -2434,7 +2433,7 @@ vect_analyze_group_access (struct data_reference * > /* Check that STEP is a multiple of type size. */ > if (dr_step && (dr_step % type_size) != 0) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "step is not a multiple of type size: step "); > @@ -2450,7 +2449,7 @@ vect_analyze_group_access (struct data_reference * > groupsize = count; > > GROUP_SIZE (vinfo_for_stmt (stmt)) = groupsize; > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Detected interleaving of size %d", (int)groupsize); > > @@ -2469,13 +2468,13 @@ vect_analyze_group_access (struct data_reference * > /* There is a gap in the end of the group. */ > if (groupsize - last_accessed_element > 0 && loop_vinfo) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Data access with gaps requires scalar " > "epilogue loop"); > if (loop->inner) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Peeling for outer loop is not supported"); > return false; > @@ -2508,7 +2507,7 @@ vect_analyze_data_ref_access (struct data_referenc > > if (loop_vinfo && !step) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "bad data-ref access in loop"); > return false; > @@ -2531,7 +2530,7 @@ vect_analyze_data_ref_access (struct data_referenc > step = STMT_VINFO_DR_STEP (stmt_info); > if (integer_zerop (step)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "zero step in outer loop."); > if (DR_IS_READ (dr)) > @@ -2557,7 +2556,7 @@ vect_analyze_data_ref_access (struct data_referenc > > if (loop && nested_in_vect_loop_p (loop, stmt)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "grouped access in outer loop."); > return false; > @@ -2588,7 +2587,7 @@ vect_analyze_data_ref_accesses (loop_vec_info loop > VEC (data_reference_p, heap) *datarefs; > struct data_reference *dr; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_analyze_data_ref_accesses ==="); > > @@ -2601,7 +2600,7 @@ vect_analyze_data_ref_accesses (loop_vec_info loop > if (STMT_VINFO_VECTORIZABLE (vinfo_for_stmt (DR_STMT (dr))) > && !vect_analyze_data_ref_access (dr)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: complicated access pattern."); > > @@ -2631,7 +2630,7 @@ vect_prune_runtime_alias_test_list (loop_vec_info > LOOP_VINFO_MAY_ALIAS_DDRS (loop_vinfo); > unsigned i, j; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_prune_runtime_alias_test_list ==="); > > @@ -2649,7 +2648,7 @@ vect_prune_runtime_alias_test_list (loop_vec_info > > if (vect_vfa_range_equal (ddr_i, ddr_j)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "found equal ranges "); > @@ -2677,7 +2676,7 @@ vect_prune_runtime_alias_test_list (loop_vec_info > if (VEC_length (ddr_p, ddrs) > > (unsigned) PARAM_VALUE (PARAM_VECT_MAX_VERSION_FOR_ALIAS_CHECKS)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "disable versioning for alias - max number of " > @@ -2964,7 +2963,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > tree scalar_type; > bool res, stop_bb_analysis = false; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_analyze_data_refs ===\n"); > > @@ -2979,7 +2978,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > > if (!res) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: loop contains function calls" > " or data references that cannot be analyzed"); > @@ -3011,7 +3010,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > if (!compute_all_dependences (BB_VINFO_DATAREFS (bb_vinfo), > &BB_VINFO_DDRS (bb_vinfo), NULL, true)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: basic block contains function" > " calls or data references that cannot be" > @@ -3035,7 +3034,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > > if (!dr || !DR_REF (dr)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: unhandled data-ref "); > return false; > @@ -3081,7 +3080,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > > if (!gather) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: data ref analysis " > @@ -3102,7 +3101,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > > if (TREE_CODE (DR_BASE_ADDRESS (dr)) == INTEGER_CST) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: base addr of dr is a " > "constant"); > @@ -3121,7 +3120,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > > if (TREE_THIS_VOLATILE (DR_REF (dr))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: volatile type "); > @@ -3140,7 +3139,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > > if (stmt_can_throw_internal (stmt)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: statement can throw an " > @@ -3163,7 +3162,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > if (TREE_CODE (DR_REF (dr)) == COMPONENT_REF > && DECL_BIT_FIELD (TREE_OPERAND (DR_REF (dr), 1))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: statement is bitfield " > @@ -3189,7 +3188,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > > if (is_gimple_call (stmt)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: dr in a call "); > @@ -3232,7 +3231,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > tree inner_base = build_fold_indirect_ref > (fold_build_pointer_plus (base, init)); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "analyze in outer-loop: "); > @@ -3245,7 +3244,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > > if (pbitpos % BITS_PER_UNIT != 0) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "failed: bit offset alignment.\n"); > return false; > @@ -3255,7 +3254,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > if (!simple_iv (loop, loop_containing_stmt (stmt), outer_base, > &base_iv, false)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "failed: evolution of base is not affine.\n"); > return false; > @@ -3278,7 +3277,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > else if (!simple_iv (loop, loop_containing_stmt (stmt), poffset, > &offset_iv, false)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "evolution of offset is not affine.\n"); > return false; > @@ -3303,7 +3302,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > STMT_VINFO_DR_ALIGNED_TO (stmt_info) = > size_int (highest_pow2_factor (offset_iv.base)); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "\touter base_address: "); > @@ -3327,7 +3326,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > > if (STMT_VINFO_DATA_REF (stmt_info)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: more than one data ref " > @@ -3355,7 +3354,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > get_vectype_for_scalar_type (scalar_type); > if (!STMT_VINFO_VECTYPE (stmt_info)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: no vectype for stmt: "); > @@ -3406,7 +3405,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > { > STMT_VINFO_DATA_REF (stmt_info) = NULL; > free_data_ref (dr); > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: not suitable for gather " > @@ -3459,7 +3458,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > > if (bad) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: data dependence conflict" > @@ -3480,7 +3479,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, > = vect_check_strided_load (stmt, loop_vinfo, NULL, NULL); > if (!strided_load) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: not suitable for strided " > @@ -3668,7 +3667,7 @@ vect_create_addr_base_for_vector_ref (gimple stmt, > mark_ptr_info_alignment_unknown (SSA_NAME_PTR_INFO (vec_stmt)); > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "created "); > dump_generic_expr (MSG_NOTE, TDF_SLIM, vec_stmt); > @@ -3790,7 +3789,7 @@ vect_create_data_ref_ptr (gimple stmt, tree aggr_t > in LOOP. */ > base_name = build_fold_indirect_ref (unshare_expr (DR_BASE_ADDRESS (dr))); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > tree data_ref_base = base_name; > dump_printf_loc (MSG_NOTE, vect_location, > @@ -4120,7 +4119,7 @@ vect_grouped_store_supported (tree vectype, unsign > /* vect_permute_store_chain requires the group size to be a power of two. */ > if (exact_log2 (count) == -1) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "the size of the group of accesses" > " is not a power of 2"); > @@ -4146,7 +4145,7 @@ vect_grouped_store_supported (tree vectype, unsign > } > } > > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf (MSG_MISSED_OPTIMIZATION, > "interleave op not supported by target."); > return false; > @@ -4564,7 +4563,7 @@ vect_grouped_load_supported (tree vectype, unsigne > /* vect_permute_load_chain requires the group size to be a power of two. */ > if (exact_log2 (count) == -1) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "the size of the group of accesses" > " is not a power of 2"); > @@ -4588,7 +4587,7 @@ vect_grouped_load_supported (tree vectype, unsigne > } > } > > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "extract even/odd not supported by target"); > return false; > Index: tree-vect-patterns.c > =================================================================== > --- tree-vect-patterns.c (revision 192695) > +++ tree-vect-patterns.c (working copy) > @@ -416,7 +416,7 @@ vect_recog_dot_prod_pattern (VEC (gimple, heap) ** > pattern_stmt = gimple_build_assign_with_ops (DOT_PROD_EXPR, var, > oprnd00, oprnd01, oprnd1); > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "vect_recog_dot_prod_pattern: detected: "); > @@ -676,7 +676,7 @@ vect_recog_widen_mult_pattern (VEC (gimple, heap) > return NULL; > > /* Pattern detected. */ > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "vect_recog_widen_mult_pattern: detected: "); > > @@ -699,7 +699,7 @@ vect_recog_widen_mult_pattern (VEC (gimple, heap) > pattern_stmt = gimple_build_assign_with_ops (WIDEN_MULT_EXPR, var, oprnd0, > oprnd1); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_gimple_stmt_loc (MSG_NOTE, vect_location, TDF_SLIM, pattern_stmt, 0); > > VEC_safe_push (gimple, heap, *stmts, last_stmt); > @@ -912,7 +912,7 @@ vect_recog_widen_sum_pattern (VEC (gimple, heap) * > pattern_stmt = gimple_build_assign_with_ops (WIDEN_SUM_EXPR, var, > oprnd0, oprnd1); > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "vect_recog_widen_sum_pattern: detected: "); > @@ -1217,7 +1217,7 @@ vect_recog_over_widening_pattern (VEC (gimple, hea > STMT_VINFO_RELATED_STMT (vinfo_for_stmt (stmt)) = pattern_stmt; > new_pattern_def_seq (vinfo_for_stmt (stmt), new_def_stmt); > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "created pattern stmt: "); > @@ -1285,7 +1285,7 @@ vect_recog_over_widening_pattern (VEC (gimple, hea > return NULL; > > /* Pattern detected. */ > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "vect_recog_over_widening_pattern: detected: "); > @@ -1421,7 +1421,7 @@ vect_recog_widen_shift_pattern (VEC (gimple, heap) > return NULL; > > /* Pattern detected. */ > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "vect_recog_widen_shift_pattern: detected: "); > > @@ -1445,7 +1445,7 @@ vect_recog_widen_shift_pattern (VEC (gimple, heap) > pattern_stmt = > gimple_build_assign_with_ops (WIDEN_LSHIFT_EXPR, var, oprnd0, oprnd1); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_gimple_stmt_loc (MSG_NOTE, vect_location, TDF_SLIM, pattern_stmt, 0); > > VEC_safe_push (gimple, heap, *stmts, last_stmt); > @@ -1567,7 +1567,7 @@ vect_recog_vector_vector_shift_pattern (VEC (gimpl > } > > /* Pattern detected. */ > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "vect_recog_vector_vector_shift_pattern: detected: "); > > @@ -1575,7 +1575,7 @@ vect_recog_vector_vector_shift_pattern (VEC (gimpl > var = vect_recog_temp_ssa_var (TREE_TYPE (oprnd0), NULL); > pattern_stmt = gimple_build_assign_with_ops (rhs_code, var, oprnd0, def); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_gimple_stmt_loc (MSG_NOTE, vect_location, TDF_SLIM, pattern_stmt, 0); > > VEC_safe_push (gimple, heap, *stmts, last_stmt); > @@ -1685,7 +1685,7 @@ vect_recog_divmod_pattern (VEC (gimple, heap) **st > return NULL; > > /* Pattern detected. */ > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "vect_recog_divmod_pattern: detected: "); > > @@ -1789,7 +1789,7 @@ vect_recog_divmod_pattern (VEC (gimple, heap) **st > signmask); > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_gimple_stmt_loc (MSG_NOTE, vect_location, TDF_SLIM, pattern_stmt, > 0); > > @@ -2031,7 +2031,7 @@ vect_recog_divmod_pattern (VEC (gimple, heap) **st > } > > /* Pattern detected. */ > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "vect_recog_divmod_pattern: detected: "); > @@ -2199,7 +2199,7 @@ vect_recog_mixed_size_cond_pattern (VEC (gimple, h > *type_in = vecitype; > *type_out = vectype; > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "vect_recog_mixed_size_cond_pattern: detected: "); > > @@ -2592,7 +2592,7 @@ vect_recog_bool_pattern (VEC (gimple, heap) **stmt > *type_out = vectype; > *type_in = vectype; > VEC_safe_push (gimple, heap, *stmts, last_stmt); > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "vect_recog_bool_pattern: detected: "); > > @@ -2638,7 +2638,7 @@ vect_recog_bool_pattern (VEC (gimple, heap) **stmt > *type_out = vectype; > *type_in = vectype; > VEC_safe_push (gimple, heap, *stmts, last_stmt); > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "vect_recog_bool_pattern: detected: "); > return pattern_stmt; > @@ -2788,7 +2788,7 @@ vect_pattern_recog_1 (vect_recog_func_ptr vect_rec > } > > /* Found a vectorizable pattern. */ > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "pattern recognized: "); > @@ -2814,7 +2814,7 @@ vect_pattern_recog_1 (vect_recog_func_ptr vect_rec > { > stmt_info = vinfo_for_stmt (stmt); > pattern_stmt = STMT_VINFO_RELATED_STMT (stmt_info); > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "additional pattern stmt: "); > @@ -2915,7 +2915,7 @@ vect_pattern_recog (loop_vec_info loop_vinfo, bb_v > VEC (gimple, heap) *stmts_to_replace = VEC_alloc (gimple, heap, 1); > gimple stmt; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_pattern_recog ==="); > > Index: tree-vect-stmts.c > =================================================================== > --- tree-vect-stmts.c (revision 192695) > +++ tree-vect-stmts.c (working copy) > @@ -190,7 +190,7 @@ vect_mark_relevant (VEC(gimple,heap) **worklist, g > bool save_live_p = STMT_VINFO_LIVE_P (stmt_info); > gimple pattern_stmt; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "mark relevant %d, live %d.", relevant, live_p); > > @@ -246,7 +246,7 @@ vect_mark_relevant (VEC(gimple,heap) **worklist, g > > pattern_stmt = STMT_VINFO_RELATED_STMT (stmt_info); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "last stmt in pattern. don't mark" > " relevant/live."); > @@ -265,7 +265,7 @@ vect_mark_relevant (VEC(gimple,heap) **worklist, g > if (STMT_VINFO_RELEVANT (stmt_info) == save_relevant > && STMT_VINFO_LIVE_P (stmt_info) == save_live_p) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "already marked relevant/live."); > return; > @@ -310,7 +310,7 @@ vect_stmt_relevant_p (gimple stmt, loop_vec_info l > if (gimple_code (stmt) != GIMPLE_PHI) > if (gimple_vdef (stmt)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vec_stmt_relevant_p: stmt has vdefs."); > *relevant = vect_used_in_scope; > @@ -324,7 +324,7 @@ vect_stmt_relevant_p (gimple stmt, loop_vec_info l > basic_block bb = gimple_bb (USE_STMT (use_p)); > if (!flow_bb_inside_loop_p (loop, bb)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vec_stmt_relevant_p: used out of loop."); > > @@ -437,7 +437,7 @@ process_use (gimple stmt, tree use, loop_vec_info > > if (!vect_is_simple_use (use, stmt, loop_vinfo, NULL, &def_stmt, &def, &dt)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: unsupported use in stmt."); > return false; > @@ -449,7 +449,7 @@ process_use (gimple stmt, tree use, loop_vec_info > def_bb = gimple_bb (def_stmt); > if (!flow_bb_inside_loop_p (loop, def_bb)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "def_stmt is out of loop."); > return true; > } > @@ -467,7 +467,7 @@ process_use (gimple stmt, tree use, loop_vec_info > && STMT_VINFO_DEF_TYPE (dstmt_vinfo) == vect_reduction_def > && bb->loop_father == def_bb->loop_father) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "reduc-stmt defining reduc-phi in the same nest."); > if (STMT_VINFO_IN_PATTERN_P (dstmt_vinfo)) > @@ -487,7 +487,7 @@ process_use (gimple stmt, tree use, loop_vec_info > ... */ > if (flow_loop_nested_p (def_bb->loop_father, bb->loop_father)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "outer-loop def-stmt defining inner-loop stmt."); > > @@ -525,7 +525,7 @@ process_use (gimple stmt, tree use, loop_vec_info > stmt # use (d) */ > else if (flow_loop_nested_p (bb->loop_father, def_bb->loop_father)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "inner-loop def-stmt defining outer-loop stmt."); > > @@ -589,7 +589,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo > enum vect_relevant relevant, tmp_relevant; > enum vect_def_type def_type; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_mark_stmts_to_be_vectorized ==="); > > @@ -602,7 +602,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo > for (si = gsi_start_phis (bb); !gsi_end_p (si); gsi_next (&si)) > { > phi = gsi_stmt (si); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "init: phi relevant? "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0); > @@ -614,7 +614,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo > for (si = gsi_start_bb (bb); !gsi_end_p (si); gsi_next (&si)) > { > stmt = gsi_stmt (si); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "init: stmt relevant? "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, stmt, 0); > @@ -632,7 +632,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo > ssa_op_iter iter; > > stmt = VEC_pop (gimple, worklist); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "worklist: > examine stmt: "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, stmt, 0); > @@ -677,7 +677,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo > /* fall through */ > > default: > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "unsupported use of reduction."); > VEC_free (gimple, heap, worklist); > @@ -692,7 +692,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo > && tmp_relevant != vect_used_in_outer_by_reduction > && tmp_relevant != vect_used_in_outer) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "unsupported use of nested cycle."); > > @@ -707,7 +707,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo > if (tmp_relevant != vect_unused_in_scope > && tmp_relevant != vect_used_by_reduction) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "unsupported use of double reduction."); > > @@ -830,7 +830,7 @@ vect_model_simple_cost (stmt_vec_info stmt_info, i > inside_cost = record_stmt_cost (body_cost_vec, ncopies, vector_stmt, > stmt_info, 0, vect_body); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_simple_cost: inside_cost = %d, " > "prologue_cost = %d .", inside_cost, prologue_cost); > @@ -876,7 +876,7 @@ vect_model_promotion_demotion_cost (stmt_vec_info > prologue_cost += add_stmt_cost (target_cost_data, 1, vector_stmt, > stmt_info, 0, vect_prologue); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_promotion_demotion_cost: inside_cost = %d, " > "prologue_cost = %d .", inside_cost, prologue_cost); > @@ -960,7 +960,7 @@ vect_model_store_cost (stmt_vec_info stmt_info, in > inside_cost = record_stmt_cost (body_cost_vec, nstmts, vec_perm, > stmt_info, 0, vect_body); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_store_cost: strided group_size = %d .", > group_size); > @@ -969,7 +969,7 @@ vect_model_store_cost (stmt_vec_info stmt_info, in > /* Costs of the stores. */ > vect_get_store_cost (first_dr, ncopies, &inside_cost, body_cost_vec); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_store_cost: inside_cost = %d, " > "prologue_cost = %d .", inside_cost, prologue_cost); > @@ -994,7 +994,7 @@ vect_get_store_cost (struct data_reference *dr, in > vector_store, stmt_info, 0, > vect_body); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_store_cost: aligned."); > break; > @@ -1006,7 +1006,7 @@ vect_get_store_cost (struct data_reference *dr, in > *inside_cost += record_stmt_cost (body_cost_vec, ncopies, > unaligned_store, stmt_info, > DR_MISALIGNMENT (dr), vect_body); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_store_cost: unaligned supported by " > "hardware."); > @@ -1017,7 +1017,7 @@ vect_get_store_cost (struct data_reference *dr, in > { > *inside_cost = VECT_MAX_COST; > > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "vect_model_store_cost: unsupported access."); > break; > @@ -1076,7 +1076,7 @@ vect_model_load_cost (stmt_vec_info stmt_info, int > inside_cost += record_stmt_cost (body_cost_vec, nstmts, vec_perm, > stmt_info, 0, vect_body); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_load_cost: strided group_size = %d .", > group_size); > @@ -1100,7 +1100,7 @@ vect_model_load_cost (stmt_vec_info stmt_info, int > &inside_cost, &prologue_cost, > prologue_cost_vec, body_cost_vec, true); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_load_cost: inside_cost = %d, " > "prologue_cost = %d .", inside_cost, prologue_cost); > @@ -1127,7 +1127,7 @@ vect_get_load_cost (struct data_reference *dr, int > *inside_cost += record_stmt_cost (body_cost_vec, ncopies, vector_load, > stmt_info, 0, vect_body); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_load_cost: aligned."); > > @@ -1140,7 +1140,7 @@ vect_get_load_cost (struct data_reference *dr, int > unaligned_load, stmt_info, > DR_MISALIGNMENT (dr), vect_body); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_load_cost: unaligned supported by " > "hardware."); > @@ -1161,7 +1161,7 @@ vect_get_load_cost (struct data_reference *dr, int > *inside_cost += record_stmt_cost (body_cost_vec, 1, vector_stmt, > stmt_info, 0, vect_body); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_load_cost: explicit realign"); > > @@ -1169,7 +1169,7 @@ vect_get_load_cost (struct data_reference *dr, int > } > case dr_explicit_realign_optimized: > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_load_cost: unaligned software " > "pipelined."); > @@ -1197,7 +1197,7 @@ vect_get_load_cost (struct data_reference *dr, int > *inside_cost += record_stmt_cost (body_cost_vec, ncopies, vec_perm, > stmt_info, 0, vect_body); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vect_model_load_cost: explicit realign optimized"); > > @@ -1208,7 +1208,7 @@ vect_get_load_cost (struct data_reference *dr, int > { > *inside_cost = VECT_MAX_COST; > > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "vect_model_load_cost: unsupported access."); > break; > @@ -1258,7 +1258,7 @@ vect_init_vector_1 (gimple stmt, gimple new_stmt, > } > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "created new init_stmt: "); > @@ -1340,7 +1340,7 @@ vect_get_vec_def_for_operand (tree op, gimple stmt > bool is_simple_use; > tree vector_type; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "vect_get_vec_def_for_operand: "); > @@ -1350,7 +1350,7 @@ vect_get_vec_def_for_operand (tree op, gimple stmt > is_simple_use = vect_is_simple_use (op, stmt, loop_vinfo, NULL, > &def_stmt, &def, &dt); > gcc_assert (is_simple_use); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > int loc_printed = 0; > if (def) > @@ -1382,7 +1382,7 @@ vect_get_vec_def_for_operand (tree op, gimple stmt > *scalar_def = op; > > /* Create 'vect_cst_ = {cst,cst,...,cst}' */ > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Create vector_cst. nunits = %d", nunits); > > @@ -1399,7 +1399,7 @@ vect_get_vec_def_for_operand (tree op, gimple stmt > *scalar_def = def; > > /* Create 'vec_inv = {inv,inv,..,inv}' */ > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "Create vector_inv."); > > return vect_init_vector (stmt, def, vector_type, NULL); > @@ -1661,7 +1661,7 @@ vect_finish_stmt_generation (gimple stmt, gimple v > set_vinfo_for_stmt (vec_stmt, new_stmt_vec_info (vec_stmt, loop_vinfo, > bb_vinfo)); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "add new stmt: "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, vec_stmt, 0); > @@ -1764,7 +1764,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat > if (rhs_type > && !types_compatible_p (rhs_type, TREE_TYPE (op))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "argument types differ."); > return false; > @@ -1775,7 +1775,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat > if (!vect_is_simple_use_1 (op, stmt, loop_vinfo, bb_vinfo, > &def_stmt, &def, &dt[i], &opvectype)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "use not simple."); > return false; > @@ -1786,7 +1786,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat > else if (opvectype > && opvectype != vectype_in) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "argument vector types differ."); > return false; > @@ -1800,7 +1800,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat > gcc_assert (vectype_in); > if (!vectype_in) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "no vectype for scalar type "); > @@ -1829,7 +1829,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat > fndecl = vectorizable_function (stmt, vectype_out, vectype_in); > if (fndecl == NULL_TREE) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "function is not vectorizable."); > > @@ -1852,7 +1852,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat > if (!vec_stmt) /* transformation not required. */ > { > STMT_VINFO_TYPE (stmt_info) = call_vec_info_type; > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "=== vectorizable_call ==="); > vect_model_simple_cost (stmt_info, ncopies, dt, NULL, NULL); > return true; > @@ -1860,7 +1860,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat > > /** Transform. **/ > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "transform call."); > > /* Handle def. */ > @@ -2375,7 +2375,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ > && (TYPE_PRECISION (rhs_type) > != GET_MODE_PRECISION (TYPE_MODE (rhs_type))))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "type conversion to/from bit-precision unsupported."); > return false; > @@ -2385,7 +2385,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ > if (!vect_is_simple_use_1 (op0, stmt, loop_vinfo, bb_vinfo, > &def_stmt, &def, &dt[0], &vectype_in)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "use not simple."); > return false; > @@ -2407,7 +2407,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ > > if (!ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "use not simple."); > return false; > @@ -2422,7 +2422,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ > gcc_assert (vectype_in); > if (!vectype_in) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "no vectype for scalar type "); > @@ -2466,7 +2466,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ > break; > /* FALLTHRU */ > unsupported: > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "conversion not supported by target."); > return false; > @@ -2565,7 +2565,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ > > if (!vec_stmt) /* transformation not required. */ > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vectorizable_conversion ==="); > if (code == FIX_TRUNC_EXPR || code == FLOAT_EXPR) > @@ -2588,7 +2588,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ > } > > /** Transform. **/ > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "transform conversion. ncopies = %d.", ncopies); > > @@ -2941,7 +2941,7 @@ vectorizable_assignment (gimple stmt, gimple_stmt_ > if (!vect_is_simple_use_1 (op, stmt, loop_vinfo, bb_vinfo, > &def_stmt, &def, &dt[0], &vectype_in)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "use not simple."); > return false; > @@ -2970,7 +2970,7 @@ vectorizable_assignment (gimple stmt, gimple_stmt_ > > TYPE_PRECISION (TREE_TYPE (op))) > && TYPE_UNSIGNED (TREE_TYPE (op)))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "type conversion to/from bit-precision " > "unsupported."); > @@ -2980,7 +2980,7 @@ vectorizable_assignment (gimple stmt, gimple_stmt_ > if (!vec_stmt) /* transformation not required. */ > { > STMT_VINFO_TYPE (stmt_info) = assignment_vec_info_type; > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vectorizable_assignment ==="); > vect_model_simple_cost (stmt_info, ncopies, dt, NULL, NULL); > @@ -2988,7 +2988,7 @@ vectorizable_assignment (gimple stmt, gimple_stmt_ > } > > /** Transform. **/ > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "transform assignment."); > > /* Handle def. */ > @@ -3135,7 +3135,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > if (TYPE_PRECISION (TREE_TYPE (scalar_dest)) > != GET_MODE_PRECISION (TYPE_MODE (TREE_TYPE (scalar_dest)))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "bit-precision shifts not supported."); > return false; > @@ -3145,7 +3145,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > if (!vect_is_simple_use_1 (op0, stmt, loop_vinfo, bb_vinfo, > &def_stmt, &def, &dt[0], &vectype)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "use not simple."); > return false; > @@ -3158,7 +3158,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > gcc_assert (vectype); > if (!vectype) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "no vectype for scalar type "); > return false; > @@ -3173,7 +3173,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > if (!vect_is_simple_use_1 (op1, stmt, loop_vinfo, bb_vinfo, &def_stmt, > &def, &dt[1], &op1_vectype)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "use not simple."); > return false; > @@ -3218,7 +3218,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > } > else > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "operand mode requires invariant argument."); > return false; > @@ -3228,7 +3228,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > if (!scalar_shift_arg) > { > optab = optab_for_tree_code (code, vectype, optab_vector); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vector/vector shift/rotate found."); > > @@ -3237,7 +3237,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > if (op1_vectype == NULL_TREE > || TYPE_MODE (op1_vectype) != TYPE_MODE (vectype)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "unusable type for last operand in" > " vector/vector shift/rotate."); > @@ -3252,7 +3252,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > if (optab > && optab_handler (optab, TYPE_MODE (vectype)) != CODE_FOR_nothing) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vector/scalar shift/rotate found."); > } > @@ -3265,7 +3265,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > { > scalar_shift_arg = false; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vector/vector shift/rotate found."); > > @@ -3282,7 +3282,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > && TYPE_MODE (TREE_TYPE (vectype)) > != TYPE_MODE (TREE_TYPE (op1))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, > vect_location, > "unusable type for last operand in" > " vector/vector shift/rotate."); > @@ -3302,7 +3302,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > /* Supportable by target? */ > if (!optab) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "no optab."); > return false; > @@ -3311,7 +3311,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > icode = (int) optab_handler (optab, vec_mode); > if (icode == CODE_FOR_nothing) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "op not supported by target."); > /* Check only during analysis. */ > @@ -3319,7 +3319,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > || (vf < vect_min_worthwhile_factor (code) > && !vec_stmt)) > return false; > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "proceeding using > word mode."); > } > > @@ -3328,7 +3328,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > && vf < vect_min_worthwhile_factor (code) > && !vec_stmt) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not worthwhile without SIMD support."); > return false; > @@ -3337,7 +3337,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > if (!vec_stmt) /* transformation not required. */ > { > STMT_VINFO_TYPE (stmt_info) = shift_vec_info_type; > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "=== > vectorizable_shift ==="); > vect_model_simple_cost (stmt_info, ncopies, dt, NULL, NULL); > return true; > @@ -3345,7 +3345,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > > /** Transform. **/ > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "transform binary/unary operation."); > > @@ -3382,7 +3382,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera > optab_op2_mode = insn_data[icode].operand[2].mode; > if (!VECTOR_MODE_P (optab_op2_mode)) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "operand 1 using scalar mode."); > vec_oprnd1 = op1; > @@ -3510,7 +3510,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i > op_type = TREE_CODE_LENGTH (code); > if (op_type != unary_op && op_type != binary_op && op_type != ternary_op) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "num. args = %d (not unary/binary/ternary op).", > op_type); > @@ -3529,7 +3529,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i > && code != BIT_XOR_EXPR > && code != BIT_AND_EXPR) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "bit-precision arithmetic not supported."); > return false; > @@ -3539,7 +3539,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i > if (!vect_is_simple_use_1 (op0, stmt, loop_vinfo, bb_vinfo, > &def_stmt, &def, &dt[0], &vectype)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "use not simple."); > return false; > @@ -3552,7 +3552,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i > gcc_assert (vectype); > if (!vectype) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "no vectype for scalar type "); > @@ -3574,7 +3574,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i > if (!vect_is_simple_use (op1, stmt, loop_vinfo, bb_vinfo, &def_stmt, > &def, &dt[1])) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "use not simple."); > return false; > @@ -3586,7 +3586,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i > if (!vect_is_simple_use (op2, stmt, loop_vinfo, bb_vinfo, &def_stmt, > &def, &dt[2])) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "use not simple."); > return false; > @@ -3628,7 +3628,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i > optab = optab_for_tree_code (code, vectype, optab_default); > if (!optab) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "no optab."); > return false; > @@ -3638,14 +3638,14 @@ vectorizable_operation (gimple stmt, gimple_stmt_i > > if (icode == CODE_FOR_nothing) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "op not supported by target."); > /* Check only during analysis. */ > if (GET_MODE_SIZE (vec_mode) != UNITS_PER_WORD > || (!vec_stmt && vf < vect_min_worthwhile_factor (code))) > return false; > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "proceeding using word mode."); > } > > @@ -3654,7 +3654,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i > && !vec_stmt > && vf < vect_min_worthwhile_factor (code)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not worthwhile without SIMD support."); > return false; > @@ -3663,7 +3663,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i > if (!vec_stmt) /* transformation not required. */ > { > STMT_VINFO_TYPE (stmt_info) = op_vec_info_type; > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vectorizable_operation ==="); > vect_model_simple_cost (stmt_info, ncopies, dt, NULL, NULL); > @@ -3672,7 +3672,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i > > /** Transform. **/ > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "transform binary/unary operation."); > > @@ -3860,7 +3860,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera > /* FORNOW. This restriction should be relaxed. */ > if (loop && nested_in_vect_loop_p (loop, stmt) && ncopies > 1) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "multiple types in nested loop."); > return false; > @@ -3894,7 +3894,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera > if (!vect_is_simple_use (op, stmt, loop_vinfo, bb_vinfo, &def_stmt, > &def, &dt)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "use not simple."); > return false; > @@ -3915,7 +3915,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera > ? STMT_VINFO_DR_STEP (stmt_info) : DR_STEP (dr), > size_zero_node) < 0) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "negative step for store."); > return false; > @@ -3946,7 +3946,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera > if (!vect_is_simple_use (op, next_stmt, loop_vinfo, bb_vinfo, > &def_stmt, &def, &dt)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "use not simple."); > return false; > @@ -4008,7 +4008,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera > group_size = vec_num = 1; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "transform store. ncopies = %d", ncopies); > > @@ -4396,7 +4396,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat > /* FORNOW. This restriction should be relaxed. */ > if (nested_in_vect_loop && ncopies > 1) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "multiple types in nested loop."); > return false; > @@ -4436,7 +4436,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat > (e.g. - data copies). */ > if (optab_handler (mov_optab, mode) == CODE_FOR_nothing) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Aligned load, but unsupported type."); > return false; > @@ -4472,7 +4472,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat > &def_stmt, &def, &gather_dt, > &gather_off_vectype)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "gather index use not simple."); > return false; > @@ -4492,7 +4492,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat > size_zero_node) < 0; > if (negative && ncopies > 1) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "multiple types with negative step."); > return false; > @@ -4505,14 +4505,14 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat > if (alignment_support_scheme != dr_aligned > && alignment_support_scheme != dr_unaligned_supported) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "negative step but alignment required."); > return false; > } > if (!perm_mask_for_reverse (vectype)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "negative step and reversing not supported."); > return false; > @@ -4527,7 +4527,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat > return true; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "transform load. ncopies = %d", ncopies); > > @@ -5334,7 +5334,7 @@ vectorizable_condition (gimple stmt, gimple_stmt_i > /* FORNOW: not yet supported. */ > if (STMT_VINFO_LIVE_P (stmt_info)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "value used after loop."); > return false; > @@ -5534,7 +5534,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect > gimple pattern_stmt; > gimple_seq pattern_def_seq; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "==> examining statement: "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, stmt, 0); > @@ -5542,7 +5542,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect > > if (gimple_has_volatile_ops (stmt)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: stmt has volatile operands"); > > @@ -5575,7 +5575,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect > /* Analyze PATTERN_STMT instead of the original stmt. */ > stmt = pattern_stmt; > stmt_info = vinfo_for_stmt (pattern_stmt); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "==> examining pattern statement: "); > @@ -5584,7 +5584,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect > } > else > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "irrelevant."); > > return true; > @@ -5597,7 +5597,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect > || STMT_VINFO_LIVE_P (vinfo_for_stmt (pattern_stmt)))) > { > /* Analyze PATTERN_STMT too. */ > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "==> examining pattern statement: "); > @@ -5621,7 +5621,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect > || STMT_VINFO_LIVE_P (vinfo_for_stmt (pattern_def_stmt))) > { > /* Analyze def stmt of STMT if it's a pattern stmt. */ > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "==> examining pattern def statement: "); > @@ -5660,7 +5660,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect > gcc_assert (PURE_SLP_STMT (stmt_info)); > > scalar_type = TREE_TYPE (gimple_get_lhs (stmt)); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "get vectype for scalar type: "); > @@ -5670,7 +5670,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect > vectype = get_vectype_for_scalar_type (scalar_type); > if (!vectype) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not SLPed: unsupported data-type "); > @@ -5680,7 +5680,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect > return false; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "vectype: "); > dump_generic_expr (MSG_NOTE, TDF_SLIM, vectype); > @@ -5724,7 +5724,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect > > if (!ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: relevant stmt not "); > @@ -5746,7 +5746,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect > > if (!ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: live stmt not "); > @@ -5846,7 +5846,7 @@ vect_transform_stmt (gimple stmt, gimple_stmt_iter > default: > if (!STMT_VINFO_LIVE_P (stmt_info)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "stmt not supported."); > gcc_unreachable (); > @@ -5871,7 +5871,7 @@ vect_transform_stmt (gimple stmt, gimple_stmt_iter > tree scalar_dest; > gimple exit_phi; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Record the vdef for outer-loop vectorization."); > > @@ -6108,7 +6108,7 @@ get_vectype_for_scalar_type_and_size (tree scalar_ > return NULL_TREE; > > vectype = build_vector_type (scalar_type, nunits); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "get vectype with %d units of type ", nunits); > @@ -6118,7 +6118,7 @@ get_vectype_for_scalar_type_and_size (tree scalar_ > if (!vectype) > return NULL_TREE; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "vectype: "); > dump_generic_expr (MSG_NOTE, TDF_SLIM, vectype); > @@ -6127,7 +6127,7 @@ get_vectype_for_scalar_type_and_size (tree scalar_ > if (!VECTOR_MODE_P (TYPE_MODE (vectype)) > && !INTEGRAL_MODE_P (TYPE_MODE (vectype))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "mode not supported by target."); > return NULL_TREE; > @@ -6198,7 +6198,7 @@ vect_is_simple_use (tree operand, gimple stmt, loo > *def_stmt = NULL; > *def = NULL_TREE; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "vect_is_simple_use: operand "); > @@ -6220,14 +6220,14 @@ vect_is_simple_use (tree operand, gimple stmt, loo > > if (TREE_CODE (operand) == PAREN_EXPR) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "non-associatable copy."); > operand = TREE_OPERAND (operand, 0); > } > > if (TREE_CODE (operand) != SSA_NAME) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not ssa-name."); > return false; > @@ -6236,13 +6236,13 @@ vect_is_simple_use (tree operand, gimple stmt, loo > *def_stmt = SSA_NAME_DEF_STMT (operand); > if (*def_stmt == NULL) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "no def_stmt."); > return false; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "def_stmt: "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, *def_stmt, 0); > @@ -6274,13 +6274,13 @@ vect_is_simple_use (tree operand, gimple stmt, loo > && *dt == vect_double_reduction_def > && gimple_code (stmt) != GIMPLE_PHI)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Unsupported pattern."); > return false; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "type of def: %d.", *dt); > > switch (gimple_code (*def_stmt)) > @@ -6299,7 +6299,7 @@ vect_is_simple_use (tree operand, gimple stmt, loo > break; > /* FALLTHRU */ > default: > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "unsupported defining stmt: "); > return false; > Index: tree-vect-slp.c > =================================================================== > --- tree-vect-slp.c (revision 192695) > +++ tree-vect-slp.c (working copy) > @@ -238,7 +238,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi > &def, &dt) > || (!def_stmt && dt != vect_constant_def)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: can't find def for "); > @@ -263,7 +263,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi > pattern = true; > if (!first && !oprnd_info->first_pattern) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: some of the stmts" > @@ -279,7 +279,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi > > if (dt == vect_unknown_def_type) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Unsupported pattern."); > return false; > @@ -296,7 +296,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi > break; > > default: > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "unsupported defining stmt: "); > return false; > @@ -361,7 +361,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi > { > if (number_of_oprnds != 2) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: different types "); > > @@ -388,7 +388,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi > && !types_compatible_p (oprnd_info->first_def_type, > TREE_TYPE (def_op0)))) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "Swapping operands of "); > @@ -400,7 +400,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi > } > else > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: different types "); > > @@ -435,7 +435,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi > > default: > /* FORNOW: Not supported. */ > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: illegal type of def "); > @@ -504,7 +504,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > /* For every stmt in NODE find its def stmt/s. */ > FOR_EACH_VEC_ELT (gimple, stmts, i, stmt) > { > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "Build SLP for "); > dump_gimple_stmt (MSG_NOTE, TDF_SLIM, stmt, 0); > @@ -513,7 +513,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > /* Fail to vectorize statements marked as unvectorizable. */ > if (!STMT_VINFO_VECTORIZABLE (vinfo_for_stmt (stmt))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: unvectorizable statement "); > @@ -527,7 +527,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > lhs = gimple_get_lhs (stmt); > if (lhs == NULL_TREE) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: not GIMPLE_ASSIGN nor " > @@ -544,7 +544,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > && (cond = gimple_assign_rhs1 (stmt)) > && !COMPARISON_CLASS_P (cond)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: condition is not " > @@ -560,7 +560,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > vectype = get_vectype_for_scalar_type (scalar_type); > if (!vectype) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: unsupported data-type "); > @@ -591,7 +591,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > || !gimple_call_nothrow_p (stmt) > || gimple_call_chain (stmt)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: unsupported call type "); > @@ -631,7 +631,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > > if (!optab) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: no optab."); > vect_free_oprnd_info (&oprnds_info); > @@ -640,7 +640,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > icode = (int) optab_handler (optab, vec_mode); > if (icode == CODE_FOR_nothing) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: " > "op not supported by target."); > @@ -674,7 +674,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > || first_stmt_code == COMPONENT_REF > || first_stmt_code == MEM_REF))) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: different operation " > @@ -689,7 +689,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > if (need_same_oprnds > && !operand_equal_p (first_op1, gimple_assign_rhs2 (stmt), 0)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: different shift " > @@ -710,7 +710,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > || gimple_call_fntype (first_stmt) > != gimple_call_fntype (stmt)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: different calls in "); > @@ -749,7 +749,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > || (GROUP_FIRST_ELEMENT (vinfo_for_stmt (stmt)) != stmt > && GROUP_GAP (vinfo_for_stmt (stmt)) != 1)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: grouped " > @@ -767,7 +767,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > if (loop_vinfo > && GROUP_SIZE (vinfo_for_stmt (stmt)) > ncopies * group_size) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: the number " > @@ -792,7 +792,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > && rhs_code != REALPART_EXPR > && rhs_code != IMAGPART_EXPR) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, > vect_location, > @@ -817,7 +817,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > if (vect_supportable_dr_alignment (first_dr, false) > == dr_unaligned_unsupported) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, > vect_location, > @@ -857,7 +857,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > if (TREE_CODE_CLASS (rhs_code) == tcc_reference) > { > /* Not grouped load. */ > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: not grouped load "); > @@ -875,7 +875,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > && rhs_code != COND_EXPR > && rhs_code != CALL_EXPR) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: operation"); > @@ -895,7 +895,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ > first_cond_code = TREE_CODE (cond_expr); > else if (first_cond_code != TREE_CODE (cond_expr)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: different" > @@ -1080,7 +1080,7 @@ vect_supported_slp_permutation_p (slp_instance ins > /* Check that the loads are all in the same interleaving chain. */ > if (GROUP_FIRST_ELEMENT (vinfo_for_stmt (scalar_stmt)) != first_load) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: unsupported data " > @@ -1169,7 +1169,7 @@ vect_supported_load_permutation_p (slp_instance sl > if (!slp_instn) > return false; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "Load permutation "); > FOR_EACH_VEC_ELT (int, load_permutation, i, next) > @@ -1376,7 +1376,7 @@ vect_supported_load_permutation_p (slp_instance sl > if (vect_supportable_dr_alignment (dr, false) > == dr_unaligned_unsupported) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, > vect_location, > @@ -1536,7 +1536,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf > > if (!vectype) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: unsupported data-type "); > @@ -1556,7 +1556,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf > unrolling_factor = least_common_multiple (nunits, group_size) / group_size; > if (unrolling_factor != 1 && !loop_vinfo) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: unrolling required in basic" > " block SLP"); > @@ -1618,7 +1618,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf > > if (unrolling_factor != 1 && !loop_vinfo) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: unrolling required in basic" > " block SLP"); > @@ -1645,7 +1645,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf > if (!vect_supported_load_permutation_p (new_instance, group_size, > load_permutation)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Build SLP failed: unsupported load " > @@ -1685,7 +1685,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf > VEC_safe_push (slp_instance, heap, BB_VINFO_SLP_INSTANCES (bb_vinfo), > new_instance); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > vect_print_slp_tree (MSG_NOTE, node); > > return true; > @@ -1717,7 +1717,7 @@ vect_analyze_slp (loop_vec_info loop_vinfo, bb_vec > gimple first_element; > bool ok = false; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "=== vect_analyze_slp ==="); > > if (loop_vinfo) > @@ -1736,7 +1736,7 @@ vect_analyze_slp (loop_vec_info loop_vinfo, bb_vec > > if (bb_vinfo && !ok) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "Failed to SLP the basic block."); > > @@ -1780,7 +1780,7 @@ vect_make_slp_decision (loop_vec_info loop_vinfo) > slp_instance instance; > int decided_to_slp = 0; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "=== > vect_make_slp_decision ==="); > > FOR_EACH_VEC_ELT (slp_instance, slp_instances, i, instance) > @@ -1798,7 +1798,7 @@ vect_make_slp_decision (loop_vec_info loop_vinfo) > > LOOP_VINFO_SLP_UNROLLING_FACTOR (loop_vinfo) = unrolling_factor; > > - if (decided_to_slp && dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (decided_to_slp && dump_enabled_p ()) > dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, > "Decided to SLP %d instances. Unrolling factor %d", > decided_to_slp, unrolling_factor); > @@ -1863,7 +1863,7 @@ vect_detect_hybrid_slp (loop_vec_info loop_vinfo) > VEC (slp_instance, heap) *slp_instances = LOOP_VINFO_SLP_INSTANCES > (loop_vinfo); > slp_instance instance; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "=== > vect_detect_hybrid_slp ==="); > > FOR_EACH_VEC_ELT (slp_instance, slp_instances, i, instance) > @@ -2060,7 +2060,7 @@ vect_bb_vectorization_profitable_p (bb_vec_info bb > > vec_outside_cost = vec_prologue_cost + vec_epilogue_cost; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, "Cost model analysis: \n"); > dump_printf (MSG_NOTE, " Vector inside of basic block cost: %d\n", > @@ -2097,7 +2097,7 @@ vect_slp_analyze_bb_1 (basic_block bb) > > if (!vect_analyze_data_refs (NULL, bb_vinfo, &min_vf)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: unhandled data-ref in basic " > "block.\n"); > @@ -2109,7 +2109,7 @@ vect_slp_analyze_bb_1 (basic_block bb) > ddrs = BB_VINFO_DDRS (bb_vinfo); > if (!VEC_length (ddr_p, ddrs)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: not enough data-refs in " > "basic block.\n"); > @@ -2123,7 +2123,7 @@ vect_slp_analyze_bb_1 (basic_block bb) > if (!vect_analyze_data_ref_dependences (NULL, bb_vinfo, &max_vf) > || min_vf > max_vf) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: unhandled data dependence " > "in basic block.\n"); > @@ -2134,7 +2134,7 @@ vect_slp_analyze_bb_1 (basic_block bb) > > if (!vect_analyze_data_refs_alignment (NULL, bb_vinfo)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: bad data alignment in basic " > "block.\n"); > @@ -2145,7 +2145,7 @@ vect_slp_analyze_bb_1 (basic_block bb) > > if (!vect_analyze_data_ref_accesses (NULL, bb_vinfo)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: unhandled data access in " > "basic block.\n"); > @@ -2158,7 +2158,7 @@ vect_slp_analyze_bb_1 (basic_block bb) > trees. */ > if (!vect_analyze_slp (NULL, bb_vinfo)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: failed to find SLP opportunities " > "in basic block.\n"); > @@ -2179,7 +2179,7 @@ vect_slp_analyze_bb_1 (basic_block bb) > > if (!vect_verify_datarefs_alignment (NULL, bb_vinfo)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: unsupported alignment in basic " > "block.\n"); > @@ -2189,7 +2189,7 @@ vect_slp_analyze_bb_1 (basic_block bb) > > if (!vect_slp_analyze_operations (bb_vinfo)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: bad operation in basic block.\n"); > > @@ -2201,7 +2201,7 @@ vect_slp_analyze_bb_1 (basic_block bb) > if (flag_vect_cost_model > && !vect_bb_vectorization_profitable_p (bb_vinfo)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: vectorization is not " > "profitable.\n"); > @@ -2210,7 +2210,7 @@ vect_slp_analyze_bb_1 (basic_block bb) > return NULL; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "Basic block will be vectorized using SLP\n"); > > @@ -2226,7 +2226,7 @@ vect_slp_analyze_bb (basic_block bb) > gimple_stmt_iterator gsi; > unsigned int vector_sizes; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "===vect_slp_analyze_bb===\n"); > > for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi)) > @@ -2240,7 +2240,7 @@ vect_slp_analyze_bb (basic_block bb) > > if (insns > PARAM_VALUE (PARAM_SLP_MAX_INSNS_IN_BB)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "not vectorized: too many instructions in " > "basic block.\n"); > @@ -2267,7 +2267,7 @@ vect_slp_analyze_bb (basic_block bb) > > /* Try the next biggest vector size. */ > current_vector_size = 1 << floor_log2 (vector_sizes); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "***** Re-trying analysis with " > "vector size %d\n", current_vector_size); > @@ -2292,7 +2292,7 @@ vect_update_slp_costs_according_to_vf (loop_vec_in > stmt_info_for_cost *si; > void *data = LOOP_VINFO_TARGET_COST_DATA (loop_vinfo); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "=== vect_update_slp_costs_according_to_vf ==="); > > @@ -2800,7 +2800,7 @@ vect_get_mask_element (gimple stmt, int first_mask > the next vector as well. */ > if (only_one_vec && *current_mask_element >= mask_nunits) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "permutation requires at least two vectors "); > @@ -2818,7 +2818,7 @@ vect_get_mask_element (gimple stmt, int first_mask > /* We either need the first vector too or have already moved to the > next vector. In both cases, this permutation needs three > vectors. */ > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "permutation requires at " > @@ -2884,7 +2884,7 @@ vect_transform_slp_perm_load (gimple stmt, VEC (tr > > if (!can_vec_perm_p (mode, false, NULL)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, > "no vect permute for "); > @@ -2964,7 +2964,7 @@ vect_transform_slp_perm_load (gimple stmt, VEC (tr > > if (!can_vec_perm_p (mode, false, mask)) > { > - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_MISSED_OPTIMIZATION, > vect_location, > @@ -3068,7 +3068,7 @@ vect_schedule_slp_instance (slp_tree node, slp_ins > SLP_TREE_NUMBER_OF_VEC_STMTS (node) = vec_stmts_size; > } > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE,vect_location, > "------>vectorizing SLP node starting from: "); > @@ -3177,7 +3177,7 @@ vect_schedule_slp (loop_vec_info loop_vinfo, bb_ve > /* Schedule the tree of INSTANCE. */ > is_store = vect_schedule_slp_instance (SLP_INSTANCE_TREE (instance), > instance, vf); > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, > "vectorizing stmts using SLP."); > } > @@ -3222,7 +3222,7 @@ vect_slp_transform_bb (basic_block bb) > > gcc_assert (bb_vinfo); > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > dump_printf_loc (MSG_NOTE, vect_location, "SLPing BB\n"); > > for (si = gsi_start_bb (bb); !gsi_end_p (si); gsi_next (&si)) > @@ -3230,7 +3230,7 @@ vect_slp_transform_bb (basic_block bb) > gimple stmt = gsi_stmt (si); > stmt_vec_info stmt_info; > > - if (dump_kind_p (MSG_NOTE)) > + if (dump_enabled_p ()) > { > dump_printf_loc (MSG_NOTE, vect_location, > "------>SLPing statement: "); > @@ -3248,7 +3248,7 @@ vect_slp_transform_bb (basic_block bb) > } > } > > - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) > + if (dump_enabled_p ()) > dump_printf (MSG_OPTIMIZED_LOCATIONS, "BASIC BLOCK VECTORIZED\n"); > > destroy_bb_vec_info (bb_vinfo);
Index: dumpfile.c =================================================================== --- dumpfile.c (revision 192695) +++ dumpfile.c (working copy) @@ -32,12 +32,18 @@ along with GCC; see the file COPYING3. If not see static int pflags; /* current dump_flags */ static int alt_flags; /* current opt_info flags */ -static FILE *alt_dump_file = NULL; static void dump_loc (int, FILE *, source_location); static int dump_phase_enabled_p (int); static FILE *dump_open_alternate_stream (struct dump_file_info *); +/* These are currently used for communicating between passes. + However, instead of accessing them directly, the passes can use + dump_printf () for dumps. */ +FILE *dump_file = NULL; +FILE *alt_dump_file = NULL; +const char *dump_file_name; + /* Table of tree dump switches. This must be consistent with the TREE_DUMP_INDEX enumeration in dumpfile.h. */ static struct dump_file_info dump_files[TDI_end] = @@ -514,14 +520,6 @@ dump_phase_enabled_p (int phase) } } -/* Return true if any of the dumps are enabled, false otherwise. */ - -inline bool -dump_enabled_p (void) -{ - return (dump_file || alt_dump_file); -} - /* Returns nonzero if tree dump PHASE has been initialized. */ int @@ -836,16 +834,6 @@ opt_info_switch_p (const char *arg) return opt_info_enable_all ((TDF_TREE | TDF_RTL | TDF_IPA), flags, filename); } -/* Return true if any dumps are enabled for the given MSG_TYPE, false - otherwise. */ - -bool -dump_kind_p (int msg_type) -{ - return (dump_file && (msg_type & pflags)) - || (alt_dump_file && (msg_type & alt_flags)); -} - /* Print basic block on the dump streams. */ void Index: dumpfile.h =================================================================== --- dumpfile.h (revision 192695) +++ dumpfile.h (working copy) @@ -108,7 +108,6 @@ struct dump_file_info int num; /* dump file number */ }; - /* In dumpfile.c */ extern char *get_dump_file_name (int); extern int dump_initialized_p (int); @@ -120,8 +119,6 @@ extern void dump_node (const_tree, int, FILE *); extern int dump_switch_p (const char *); extern int opt_info_switch_p (const char *); extern const char *dump_flag_name (int); -extern bool dump_kind_p (int); -extern inline bool dump_enabled_p (void); extern void dump_printf (int, const char *, ...) ATTRIBUTE_PRINTF_2; extern void dump_printf_loc (int, source_location, const char *, ...) ATTRIBUTE_PRINTF_3; @@ -142,10 +139,19 @@ extern void dump_bb (FILE *, basic_block, int, int /* Global variables used to communicate with passes. */ extern FILE *dump_file; +extern FILE *alt_dump_file; extern int dump_flags; extern const char *dump_file_name; /* Return the dump_file_info for the given phase. */ extern struct dump_file_info *get_dump_file_info (int); +/* Return true if any of the dumps are enabled, false otherwise. */ + +static inline bool +dump_enabled_p (void) +{ + return (dump_file || alt_dump_file); +} + #endif /* GCC_DUMPFILE_H */ Index: toplev.c =================================================================== --- toplev.c (revision 192695) +++ toplev.c (working copy) @@ -169,8 +169,6 @@ const char *user_label_prefix; FILE *asm_out_file; FILE *aux_info_file; FILE *stack_usage_file = NULL; -FILE *dump_file = NULL; -const char *dump_file_name; /* The current working directory of a translation. It's generally the directory from which compilation was initiated, but a preprocessed Index: tree-vect-loop-manip.c =================================================================== --- tree-vect-loop-manip.c (revision 192695) +++ tree-vect-loop-manip.c (working copy) @@ -792,7 +792,7 @@ slpeel_make_loop_iterate_ntimes (struct loop *loop free_stmt_vec_info (orig_cond); loop_loc = find_loop_location (loop); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { if (LOCATION_LOCUS (loop_loc) != UNKNOWN_LOC) dump_printf (MSG_NOTE, "\nloop at %s:%d: ", LOC_FILE (loop_loc), @@ -1683,7 +1683,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo) /* Analyze phi functions of the loop header. */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_can_advance_ivs_p:"); for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi)) { @@ -1691,7 +1691,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo) tree evolution_part; phi = gsi_stmt (gsi); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Analyze phi: "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0); @@ -1702,7 +1702,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo) if (virtual_operand_p (PHI_RESULT (phi))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "virtual phi. skip."); continue; @@ -1712,7 +1712,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo) if (STMT_VINFO_DEF_TYPE (vinfo_for_stmt (phi)) == vect_reduction_def) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "reduc phi. skip."); continue; @@ -1725,13 +1725,13 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo) if (!access_fn) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "No Access function."); return false; } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Access function of PHI: "); @@ -1742,7 +1742,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo) if (evolution_part == NULL_TREE) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf (MSG_MISSED_OPTIMIZATION, "No evolution."); return false; } @@ -1827,7 +1827,7 @@ vect_update_ivs_after_vectorizer (loop_vec_info lo phi = gsi_stmt (gsi); phi1 = gsi_stmt (gsi1); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "vect_update_ivs_after_vectorizer: phi: "); @@ -1837,7 +1837,7 @@ vect_update_ivs_after_vectorizer (loop_vec_info lo /* Skip virtual phi's. */ if (virtual_operand_p (PHI_RESULT (phi))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "virtual phi. skip."); continue; @@ -1847,7 +1847,7 @@ vect_update_ivs_after_vectorizer (loop_vec_info lo stmt_info = vinfo_for_stmt (phi); if (STMT_VINFO_DEF_TYPE (stmt_info) == vect_reduction_def) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "reduc phi. skip."); continue; @@ -1910,7 +1910,7 @@ vect_do_peeling_for_loop_bound (loop_vec_info loop tree cond_expr = NULL_TREE; gimple_seq cond_expr_stmt_list = NULL; - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "=== vect_do_peeling_for_loop_bound ==="); @@ -2022,7 +2022,7 @@ vect_gen_niters_for_prolog_loop (loop_vec_info loo { int npeel = LOOP_PEELING_FOR_ALIGNMENT (loop_vinfo); - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "known peeling = %d.", npeel); @@ -2076,7 +2076,7 @@ vect_gen_niters_for_prolog_loop (loop_vec_info loo if (TREE_CODE (loop_niters) != INTEGER_CST) iters = fold_build2 (MIN_EXPR, niters_type, iters, loop_niters); - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "niters for prolog loop: "); @@ -2134,7 +2134,7 @@ vect_update_inits_of_drs (loop_vec_info loop_vinfo VEC (data_reference_p, heap) *datarefs = LOOP_VINFO_DATAREFS (loop_vinfo); struct data_reference *dr; - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "=== vect_update_inits_of_dr ==="); @@ -2163,7 +2163,7 @@ vect_do_peeling_for_alignment (loop_vec_info loop_ int max_iter; int bound = 0; - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "=== vect_do_peeling_for_alignment ==="); @@ -2475,7 +2475,7 @@ vect_create_cond_for_alias_checks (loop_vec_info l segment_length_a = vect_vfa_segment_size (dr_a, length_factor); segment_length_b = vect_vfa_segment_size (dr_b, length_factor); - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "create runtime check for data references "); @@ -2506,7 +2506,7 @@ vect_create_cond_for_alias_checks (loop_vec_info l *cond_expr = part_cond_expr; } - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "created %u versioning for alias checks.\n", VEC_length (ddr_p, may_alias_ddrs)); Index: tree-vectorizer.c =================================================================== --- tree-vectorizer.c (revision 192695) +++ tree-vectorizer.c (working copy) @@ -107,7 +107,7 @@ vectorize_loops (void) loop_vec_info loop_vinfo; vect_location = find_loop_location (loop); if (LOCATION_LOCUS (vect_location) != UNKNOWN_LOC - && dump_kind_p (MSG_ALL)) + && dump_enabled_p ()) dump_printf (MSG_ALL, "\nAnalyzing loop at %s:%d\n", LOC_FILE (vect_location), LOC_LINE (vect_location)); @@ -118,7 +118,7 @@ vectorize_loops (void) continue; if (LOCATION_LOCUS (vect_location) != UNKNOWN_LOC - && dump_kind_p (MSG_ALL)) + && dump_enabled_p ()) dump_printf (MSG_ALL, "\n\nVectorizing loop at %s:%d\n", LOC_FILE (vect_location), LOC_LINE (vect_location)); vect_transform_loop (loop_vinfo); @@ -128,8 +128,8 @@ vectorize_loops (void) vect_location = UNKNOWN_LOC; statistics_counter_event (cfun, "Vectorized loops", num_vectorized_loops); - if (dump_kind_p (MSG_ALL) - || (num_vectorized_loops > 0 && dump_kind_p (MSG_ALL))) + if (dump_enabled_p () + || (num_vectorized_loops > 0 && dump_enabled_p ())) dump_printf_loc (MSG_ALL, vect_location, "vectorized %u loops in function.\n", num_vectorized_loops); @@ -170,7 +170,7 @@ execute_vect_slp (void) if (vect_slp_analyze_bb (bb)) { vect_slp_transform_bb (bb); - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "basic block vectorized using SLP\n"); } Index: tree-vect-loop.c =================================================================== --- tree-vect-loop.c (revision 192695) +++ tree-vect-loop.c (working copy) @@ -187,7 +187,7 @@ vect_determine_vectorization_factor (loop_vec_info gimple_stmt_iterator pattern_def_si = gsi_none (); bool analyze_pattern_stmt = false; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_determine_vectorization_factor ==="); @@ -199,7 +199,7 @@ vect_determine_vectorization_factor (loop_vec_info { phi = gsi_stmt (si); stmt_info = vinfo_for_stmt (phi); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "==> examining phi: "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0); @@ -212,7 +212,7 @@ vect_determine_vectorization_factor (loop_vec_info gcc_assert (!STMT_VINFO_VECTYPE (stmt_info)); scalar_type = TREE_TYPE (PHI_RESULT (phi)); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "get vectype for scalar type: "); @@ -222,7 +222,7 @@ vect_determine_vectorization_factor (loop_vec_info vectype = get_vectype_for_scalar_type (scalar_type); if (!vectype) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: unsupported " @@ -234,14 +234,14 @@ vect_determine_vectorization_factor (loop_vec_info } STMT_VINFO_VECTYPE (stmt_info) = vectype; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "vectype: "); dump_generic_expr (MSG_NOTE, TDF_SLIM, vectype); } nunits = TYPE_VECTOR_SUBPARTS (vectype); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "nunits = %d", nunits); if (!vectorization_factor @@ -261,7 +261,7 @@ vect_determine_vectorization_factor (loop_vec_info stmt_info = vinfo_for_stmt (stmt); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "==> examining statement: "); @@ -281,7 +281,7 @@ vect_determine_vectorization_factor (loop_vec_info { stmt = pattern_stmt; stmt_info = vinfo_for_stmt (pattern_stmt); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "==> examining pattern statement: "); @@ -290,7 +290,7 @@ vect_determine_vectorization_factor (loop_vec_info } else { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "skip."); gsi_next (&si); continue; @@ -330,7 +330,7 @@ vect_determine_vectorization_factor (loop_vec_info if (!gsi_end_p (pattern_def_si)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "==> examining pattern def stmt: "); @@ -353,7 +353,7 @@ vect_determine_vectorization_factor (loop_vec_info if (gimple_get_lhs (stmt) == NULL_TREE) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: irregular stmt."); @@ -365,7 +365,7 @@ vect_determine_vectorization_factor (loop_vec_info if (VECTOR_MODE_P (TYPE_MODE (gimple_expr_type (stmt)))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: vector stmt in loop:"); @@ -389,7 +389,7 @@ vect_determine_vectorization_factor (loop_vec_info { gcc_assert (!STMT_VINFO_DATA_REF (stmt_info)); scalar_type = TREE_TYPE (gimple_get_lhs (stmt)); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "get vectype for scalar type: "); @@ -398,7 +398,7 @@ vect_determine_vectorization_factor (loop_vec_info vectype = get_vectype_for_scalar_type (scalar_type); if (!vectype) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: unsupported " @@ -417,7 +417,7 @@ vect_determine_vectorization_factor (loop_vec_info support one vector size per loop). */ scalar_type = vect_get_smallest_scalar_type (stmt, &dummy, &dummy); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "get vectype for scalar type: "); @@ -426,7 +426,7 @@ vect_determine_vectorization_factor (loop_vec_info vf_vectype = get_vectype_for_scalar_type (scalar_type); if (!vf_vectype) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: unsupported data-type "); @@ -439,7 +439,7 @@ vect_determine_vectorization_factor (loop_vec_info if ((GET_MODE_SIZE (TYPE_MODE (vectype)) != GET_MODE_SIZE (TYPE_MODE (vf_vectype)))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: different sized vector " @@ -453,14 +453,14 @@ vect_determine_vectorization_factor (loop_vec_info return false; } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "vectype: "); dump_generic_expr (MSG_NOTE, TDF_SLIM, vf_vectype); } nunits = TYPE_VECTOR_SUBPARTS (vf_vectype); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "nunits = %d", nunits); if (!vectorization_factor || (nunits > vectorization_factor)) @@ -475,12 +475,12 @@ vect_determine_vectorization_factor (loop_vec_info } /* TODO: Analyze cost. Decide if worth while to vectorize. */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vectorization factor = %d", vectorization_factor); if (vectorization_factor <= 1) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: unsupported data-type"); return false; @@ -517,7 +517,7 @@ vect_is_simple_iv_evolution (unsigned loop_nb, tre step_expr = evolution_part; init_expr = unshare_expr (initial_condition_in_loop_num (access_fn, loop_nb)); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "step: "); dump_generic_expr (MSG_NOTE, TDF_SLIM, step_expr); @@ -530,7 +530,7 @@ vect_is_simple_iv_evolution (unsigned loop_nb, tre if (TREE_CODE (step_expr) != INTEGER_CST) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "step unknown."); return false; @@ -555,7 +555,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v gimple_stmt_iterator gsi; bool double_reduc; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_analyze_scalar_cycles ==="); @@ -569,7 +569,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v tree def = PHI_RESULT (phi); stmt_vec_info stmt_vinfo = vinfo_for_stmt (phi); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Analyze phi: "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0); @@ -587,7 +587,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v if (access_fn) { STRIP_NOPS (access_fn); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Access function of PHI: "); @@ -606,7 +606,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v gcc_assert (STMT_VINFO_LOOP_PHI_EVOLUTION_PART (stmt_vinfo) != NULL_TREE); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Detected induction."); STMT_VINFO_DEF_TYPE (stmt_vinfo) = vect_induction_def; } @@ -621,7 +621,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v gimple reduc_stmt; bool nested_cycle; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Analyze phi: "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0); @@ -637,7 +637,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v { if (double_reduc) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Detected double reduction."); @@ -649,7 +649,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v { if (nested_cycle) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Detected vectorizable nested cycle."); @@ -659,7 +659,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v } else { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Detected reduction."); @@ -675,7 +675,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v } } else - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Unknown def-use cycle pattern."); } @@ -737,7 +737,7 @@ vect_get_loop_niters (struct loop *loop, tree *num { tree niters; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== get_loop_niters ==="); niters = number_of_exit_cond_executions (loop); @@ -747,7 +747,7 @@ vect_get_loop_niters (struct loop *loop, tree *num { *number_of_iterations = niters; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "==> get_loop_niters:"); dump_generic_expr (MSG_NOTE, TDF_SLIM, *number_of_iterations); @@ -995,7 +995,7 @@ vect_analyze_loop_1 (struct loop *loop) { loop_vec_info loop_vinfo; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "===== analyze_loop_nest_1 ====="); @@ -1004,7 +1004,7 @@ vect_analyze_loop_1 (struct loop *loop) loop_vinfo = vect_analyze_loop_form (loop); if (!loop_vinfo) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "bad inner-loop form."); return NULL; @@ -1030,7 +1030,7 @@ vect_analyze_loop_form (struct loop *loop) tree number_of_iterations = NULL; loop_vec_info inner_loop_vinfo = NULL; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_analyze_loop_form ==="); @@ -1054,7 +1054,7 @@ vect_analyze_loop_form (struct loop *loop) if (loop->num_nodes != 2) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: control flow in loop."); return NULL; @@ -1062,7 +1062,7 @@ vect_analyze_loop_form (struct loop *loop) if (empty_block_p (loop->header)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: empty loop."); return NULL; @@ -1092,7 +1092,7 @@ vect_analyze_loop_form (struct loop *loop) if ((loop->inner)->inner || (loop->inner)->next) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: multiple nested loops."); return NULL; @@ -1102,7 +1102,7 @@ vect_analyze_loop_form (struct loop *loop) inner_loop_vinfo = vect_analyze_loop_1 (loop->inner); if (!inner_loop_vinfo) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: Bad inner loop."); return NULL; @@ -1111,7 +1111,7 @@ vect_analyze_loop_form (struct loop *loop) if (!expr_invariant_in_loop_p (loop, LOOP_VINFO_NITERS (inner_loop_vinfo))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: inner-loop count not invariant."); destroy_loop_vec_info (inner_loop_vinfo, true); @@ -1120,7 +1120,7 @@ vect_analyze_loop_form (struct loop *loop) if (loop->num_nodes != 5) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: control flow in loop."); destroy_loop_vec_info (inner_loop_vinfo, true); @@ -1136,14 +1136,14 @@ vect_analyze_loop_form (struct loop *loop) || !single_exit (innerloop) || single_exit (innerloop)->dest != EDGE_PRED (loop->latch, 0)->src) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: unsupported outerloop form."); destroy_loop_vec_info (inner_loop_vinfo, true); return NULL; } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Considering outer-loop vectorization."); } @@ -1151,7 +1151,7 @@ vect_analyze_loop_form (struct loop *loop) if (!single_exit (loop) || EDGE_COUNT (loop->header->preds) != 2) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { if (!single_exit (loop)) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, @@ -1172,7 +1172,7 @@ vect_analyze_loop_form (struct loop *loop) if (!empty_block_p (loop->latch) || !gimple_seq_empty_p (phi_nodes (loop->latch))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: unexpected loop form."); if (inner_loop_vinfo) @@ -1187,12 +1187,12 @@ vect_analyze_loop_form (struct loop *loop) if (!(e->flags & EDGE_ABNORMAL)) { split_loop_exit_edge (e); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf (MSG_NOTE, "split exit edge."); } else { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: abnormal loop exit edge."); if (inner_loop_vinfo) @@ -1204,7 +1204,7 @@ vect_analyze_loop_form (struct loop *loop) loop_cond = vect_get_loop_niters (loop, &number_of_iterations); if (!loop_cond) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: complicated exit condition."); if (inner_loop_vinfo) @@ -1214,7 +1214,7 @@ vect_analyze_loop_form (struct loop *loop) if (!number_of_iterations) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: number of iterations cannot be " "computed."); @@ -1225,7 +1225,7 @@ vect_analyze_loop_form (struct loop *loop) if (chrec_contains_undetermined (number_of_iterations)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Infinite number of iterations."); if (inner_loop_vinfo) @@ -1235,7 +1235,7 @@ vect_analyze_loop_form (struct loop *loop) if (!NITERS_KNOWN_P (number_of_iterations)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Symbolic number of iterations is "); @@ -1244,7 +1244,7 @@ vect_analyze_loop_form (struct loop *loop) } else if (TREE_INT_CST_LOW (number_of_iterations) == 0) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: number of iterations = 0."); if (inner_loop_vinfo) @@ -1292,7 +1292,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v HOST_WIDE_INT estimated_niter; int min_profitable_estimate; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_analyze_loop_operations ==="); @@ -1328,7 +1328,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v LOOP_VINFO_SLP_UNROLLING_FACTOR (loop_vinfo)); LOOP_VINFO_VECT_FACTOR (loop_vinfo) = vectorization_factor; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Updating vectorization factor to %d ", vectorization_factor); @@ -1344,7 +1344,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v ok = true; stmt_info = vinfo_for_stmt (phi); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "examining phi: "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0); @@ -1363,7 +1363,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v && STMT_VINFO_DEF_TYPE (stmt_info) != vect_double_reduction_def) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Unsupported loop-closed phi in " "outer-loop."); @@ -1405,7 +1405,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v if (STMT_VINFO_LIVE_P (stmt_info)) { /* FORNOW: not yet supported. */ - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: value used after loop."); return false; @@ -1415,7 +1415,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v && STMT_VINFO_DEF_TYPE (stmt_info) != vect_induction_def) { /* A scalar-dependence cycle that we don't support. */ - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: scalar dependence cycle."); return false; @@ -1430,7 +1430,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v if (!ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: relevant phi not " @@ -1456,18 +1456,17 @@ vect_analyze_loop_operations (loop_vec_info loop_v touching this loop. */ if (!need_to_vectorize) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "All the computation can be taken out of the loop."); - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: redundant loop. no profit to " "vectorize."); return false; } - if (LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo) - && dump_kind_p (MSG_NOTE)) + if (LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo) && dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vectorization_factor = %d, niters = " HOST_WIDE_INT_PRINT_DEC, vectorization_factor, @@ -1478,10 +1477,10 @@ vect_analyze_loop_operations (loop_vec_info loop_v || ((max_niter = max_stmt_executions_int (loop)) != -1 && (unsigned HOST_WIDE_INT) max_niter < vectorization_factor)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: iteration count too small."); - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: iteration count smaller than " "vectorization factor."); @@ -1500,10 +1499,10 @@ vect_analyze_loop_operations (loop_vec_info loop_v if (min_profitable_iters < 0) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: vectorization not profitable."); - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: vector version will never be " "profitable."); @@ -1526,10 +1525,10 @@ vect_analyze_loop_operations (loop_vec_info loop_v if (LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo) && LOOP_VINFO_INT_NITERS (loop_vinfo) <= th) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: vectorization not profitable."); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "not vectorized: iteration count smaller than user " "specified loop bound parameter or minimum profitable " @@ -1541,11 +1540,11 @@ vect_analyze_loop_operations (loop_vec_info loop_v && ((unsigned HOST_WIDE_INT) estimated_niter <= MAX (th, (unsigned)min_profitable_estimate))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: estimated iteration count too " "small."); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "not vectorized: estimated iteration count smaller " "than specified loop bound parameter or minimum " @@ -1558,18 +1557,18 @@ vect_analyze_loop_operations (loop_vec_info loop_v || LOOP_VINFO_INT_NITERS (loop_vinfo) % vectorization_factor != 0 || LOOP_PEELING_FOR_ALIGNMENT (loop_vinfo)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "epilog loop required."); if (!vect_can_advance_ivs_p (loop_vinfo)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: can't create epilog loop 1."); return false; } if (!slpeel_can_duplicate_loop_p (loop, single_exit (loop))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: can't create epilog loop 2."); return false; @@ -1602,7 +1601,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) ok = vect_analyze_data_refs (loop_vinfo, NULL, &min_vf); if (!ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "bad data references."); return false; @@ -1620,7 +1619,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) ok = vect_mark_stmts_to_be_vectorized (loop_vinfo); if (!ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "unexpected pattern."); return false; @@ -1635,7 +1634,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) if (!ok || max_vf < min_vf) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "bad data dependence."); return false; @@ -1644,14 +1643,14 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) ok = vect_determine_vectorization_factor (loop_vinfo); if (!ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "can't determine vectorization factor."); return false; } if (max_vf < LOOP_VINFO_VECT_FACTOR (loop_vinfo)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "bad data dependence."); return false; @@ -1663,7 +1662,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) ok = vect_analyze_data_refs_alignment (loop_vinfo, NULL); if (!ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "bad data alignment."); return false; @@ -1675,7 +1674,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) ok = vect_analyze_data_ref_accesses (loop_vinfo, NULL); if (!ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "bad data access."); return false; @@ -1687,7 +1686,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) ok = vect_prune_runtime_alias_test_list (loop_vinfo); if (!ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "too long list of versioning for alias " "run-time tests."); @@ -1700,7 +1699,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) ok = vect_enhance_data_refs_alignment (loop_vinfo); if (!ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "bad data alignment."); return false; @@ -1725,7 +1724,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo) ok = vect_analyze_loop_operations (loop_vinfo, slp); if (!ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "bad operation or unsupported loop bound."); return false; @@ -1749,7 +1748,7 @@ vect_analyze_loop (struct loop *loop) current_vector_size = 0; vector_sizes = targetm.vectorize.autovectorize_vector_sizes (); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "===== analyze_loop_nest ====="); @@ -1757,7 +1756,7 @@ vect_analyze_loop (struct loop *loop) && loop_vec_info_for_loop (loop_outer (loop)) && LOOP_VINFO_VECTORIZABLE_P (loop_vec_info_for_loop (loop_outer (loop)))) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "outer-loop already vectorized."); return NULL; @@ -1769,7 +1768,7 @@ vect_analyze_loop (struct loop *loop) loop_vinfo = vect_analyze_loop_form (loop); if (!loop_vinfo) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "bad loop form."); return NULL; @@ -1791,7 +1790,7 @@ vect_analyze_loop (struct loop *loop) /* Try the next biggest vector size. */ current_vector_size = 1 << floor_log2 (vector_sizes); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "***** Re-trying analysis with " "vector size %d\n", current_vector_size); @@ -2023,7 +2022,7 @@ vect_is_slp_reduction (loop_vec_info loop_info, gi == vect_internal_def && !is_loop_header_bb_p (gimple_bb (def_stmt))))) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "swapping oprnds: "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, next_stmt, 0); @@ -2125,7 +2124,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf if (!flow_bb_inside_loop_p (loop, gimple_bb (use_stmt))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "intermediate value used outside loop."); @@ -2137,7 +2136,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf nloop_uses++; if (nloop_uses > 1) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "reduction used in loop."); return NULL; @@ -2146,7 +2145,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf if (TREE_CODE (loop_arg) != SSA_NAME) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "reduction: not ssa_name: "); @@ -2158,7 +2157,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf def_stmt = SSA_NAME_DEF_STMT (loop_arg); if (!def_stmt) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "reduction: no def_stmt."); return NULL; @@ -2166,7 +2165,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf if (!is_gimple_assign (def_stmt) && gimple_code (def_stmt) != GIMPLE_PHI) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_gimple_stmt (MSG_NOTE, TDF_SLIM, def_stmt, 0); return NULL; } @@ -2194,7 +2193,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf nloop_uses++; if (nloop_uses > 1) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "reduction used in loop."); return NULL; @@ -2210,7 +2209,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf if (gimple_phi_num_args (def_stmt) != 1 || TREE_CODE (op1) != SSA_NAME) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "unsupported phi node definition."); @@ -2223,7 +2222,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf && flow_bb_inside_loop_p (loop->inner, gimple_bb (def1)) && is_gimple_assign (def1)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) report_vect_op (MSG_NOTE, def_stmt, "detected double reduction: "); @@ -2250,7 +2249,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf if (check_reduction && (!commutative_tree_code (code) || !associative_tree_code (code))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, "reduction: not commutative/associative: "); return NULL; @@ -2260,7 +2259,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf { if (code != COND_EXPR) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, "reduction: not binary operation: "); @@ -2279,7 +2278,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf if (TREE_CODE (op1) != SSA_NAME && TREE_CODE (op2) != SSA_NAME) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, "reduction: uses not ssa_names: "); @@ -2293,7 +2292,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf if (TREE_CODE (op1) != SSA_NAME && TREE_CODE (op2) != SSA_NAME) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, "reduction: uses not ssa_names: "); @@ -2311,7 +2310,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf || (op4 && TREE_CODE (op4) == SSA_NAME && !types_compatible_p (type, TREE_TYPE (op4)))) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "reduction: multiple types: operation type: "); @@ -2353,7 +2352,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf && check_reduction) { /* Changing the order of operations changes the semantics. */ - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, "reduction: unsafe fp math optimization: "); return NULL; @@ -2362,7 +2361,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf && check_reduction) { /* Changing the order of operations changes the semantics. */ - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, "reduction: unsafe int math optimization: "); return NULL; @@ -2370,7 +2369,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf else if (SAT_FIXED_POINT_TYPE_P (type) && check_reduction) { /* Changing the order of operations changes the semantics. */ - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, "reduction: unsafe fixed-point math optimization: "); return NULL; @@ -2407,7 +2406,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf if (code != COND_EXPR && ((!def1 || gimple_nop_p (def1)) && (!def2 || gimple_nop_p (def2)))) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) report_vect_op (MSG_NOTE, def_stmt, "reduction: no defs for operands: "); return NULL; } @@ -2429,7 +2428,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf == vect_internal_def && !is_loop_header_bb_p (gimple_bb (def1))))))) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) report_vect_op (MSG_NOTE, def_stmt, "detected reduction: "); return def_stmt; } @@ -2452,7 +2451,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf /* Swap operands (just for simplicity - so that the rest of the code can assume that the reduction variable is always the last (second) argument). */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) report_vect_op (MSG_NOTE, def_stmt, "detected reduction: need to swap operands: "); @@ -2464,7 +2463,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf } else { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) report_vect_op (MSG_NOTE, def_stmt, "detected reduction: "); } @@ -2474,14 +2473,14 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf /* Try to find SLP reduction chain. */ if (check_reduction && vect_is_slp_reduction (loop_info, phi, def_stmt)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) report_vect_op (MSG_NOTE, def_stmt, "reduction: detected reduction chain: "); return def_stmt; } - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt, "reduction: unknown pattern: "); @@ -2589,7 +2588,7 @@ vect_get_known_peeling_cost (loop_vec_info loop_vi if (!LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo)) { *peel_iters_epilogue = vf/2; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "cost model: epilogue peel iters set to vf/2 " "because loop iterations are unknown ."); @@ -2882,7 +2881,7 @@ vect_estimate_min_profitable_iters (loop_vec_info /* vector version will never be profitable. */ else { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "cost model: the vector iteration cost = %d " "divided by the scalar iteration cost = %d " @@ -2893,7 +2892,7 @@ vect_estimate_min_profitable_iters (loop_vec_info return; } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Cost model analysis: \n"); dump_printf (MSG_NOTE, " Vector inside of loop cost: %d\n", @@ -2925,7 +2924,7 @@ vect_estimate_min_profitable_iters (loop_vec_info then skip the vectorized loop. */ min_profitable_iters--; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, " Runtime profitability threshold = %d\n", min_profitable_iters); @@ -2950,7 +2949,7 @@ vect_estimate_min_profitable_iters (loop_vec_info } min_profitable_estimate --; min_profitable_estimate = MAX (min_profitable_estimate, min_profitable_iters); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, " Static estimate profitability threshold = %d\n", min_profitable_iters); @@ -3010,7 +3009,7 @@ vect_model_reduction_cost (stmt_vec_info stmt_info vectype = get_vectype_for_scalar_type (TREE_TYPE (reduction_op)); if (!vectype) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "unsupported data-type "); @@ -3081,7 +3080,7 @@ vect_model_reduction_cost (stmt_vec_info stmt_info } } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf (MSG_NOTE, "vect_model_reduction_cost: inside_cost = %d, " "prologue_cost = %d, epilogue_cost = %d .", inside_cost, @@ -3110,7 +3109,7 @@ vect_model_induction_cost (stmt_vec_info stmt_info prologue_cost = add_stmt_cost (target_cost_data, 2, scalar_to_vec, stmt_info, 0, vect_prologue); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_induction_cost: inside_cost = %d, " "prologue_cost = %d .", inside_cost, prologue_cost); @@ -3239,7 +3238,7 @@ get_initial_def_for_induction (gimple iv_phi) new_bb = gsi_insert_on_edge_immediate (pe, init_stmt); gcc_assert (!new_bb); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "created new init_stmt: "); @@ -3382,7 +3381,7 @@ get_initial_def_for_induction (gimple iv_phi) && !STMT_VINFO_LIVE_P (stmt_vinfo)); STMT_VINFO_VEC_STMT (stmt_vinfo) = new_stmt; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "vector of inductions after inner-loop:"); @@ -3392,7 +3391,7 @@ get_initial_def_for_induction (gimple iv_phi) } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "transform induction: created def-use cycle: "); @@ -3800,7 +3799,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap) add_phi_arg (phi, def, loop_latch_edge (loop), UNKNOWN_LOCATION); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "transform reduction: created def-use cycle: "); @@ -4001,7 +4000,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap) /*** Case 1: Create: v_out2 = reduc_expr <v_out1> */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Reduce using direct vector reduction."); @@ -4052,7 +4051,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap) Create: va = vop <va, va'> } */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Reduce using vector shifts"); @@ -4093,7 +4092,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap) Create: s = op <s, s'> // For non SLP cases } */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Reduce using scalar code. "); @@ -4184,7 +4183,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap) { tree rhs; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "extract scalar result"); @@ -4423,7 +4422,7 @@ vect_finalize_reduction: UNKNOWN_LOCATION); add_phi_arg (vect_phi, PHI_RESULT (inner_phi), loop_latch_edge (outer_loop), UNKNOWN_LOCATION); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "created double reduction phi node: "); @@ -4773,7 +4772,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i { if (!vectorizable_condition (stmt, gsi, NULL, ops[reduc_index], 0, NULL)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "unsupported condition in reduction"); @@ -4788,7 +4787,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i optab = optab_for_tree_code (code, vectype_in, optab_default); if (!optab) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "no optab."); @@ -4797,7 +4796,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i if (optab_handler (optab, vec_mode) == CODE_FOR_nothing) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf (MSG_NOTE, "op not supported by target."); if (GET_MODE_SIZE (vec_mode) != UNITS_PER_WORD @@ -4805,7 +4804,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i < vect_min_worthwhile_factor (code)) return false; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf (MSG_NOTE, "proceeding using word mode."); } @@ -4814,7 +4813,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i && LOOP_VINFO_VECT_FACTOR (loop_vinfo) < vect_min_worthwhile_factor (code)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not worthwhile without SIMD support."); @@ -4895,7 +4894,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i optab_default); if (!reduc_optab) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "no optab for reduction."); @@ -4905,7 +4904,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i if (reduc_optab && optab_handler (reduc_optab, vec_mode) == CODE_FOR_nothing) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "reduc op not supported by target."); @@ -4916,7 +4915,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i { if (!nested_cycle || double_reduc) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "no reduc code for scalar code."); @@ -4926,7 +4925,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i if (double_reduc && ncopies > 1) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "multiple types in double reduction"); @@ -4945,7 +4944,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i ops[1] = fold_convert (TREE_TYPE (ops[0]), ops[1]); else { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "invalid types in dot-prod"); @@ -4963,7 +4962,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i /** Transform. **/ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "transform reduction."); /* FORNOW: Multiple types are not supported for condition. */ @@ -5249,7 +5248,7 @@ vectorizable_induction (gimple phi, gimple_stmt_it if (ncopies > 1) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "multiple types in nested loop."); return false; @@ -5273,7 +5272,7 @@ vectorizable_induction (gimple phi, gimple_stmt_it if (!(STMT_VINFO_RELEVANT_P (exit_phi_vinfo) && !STMT_VINFO_LIVE_P (exit_phi_vinfo))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "inner-loop induction only used outside " "of the outer vectorized loop."); @@ -5297,7 +5296,7 @@ vectorizable_induction (gimple phi, gimple_stmt_it if (!vec_stmt) /* transformation not required. */ { STMT_VINFO_TYPE (stmt_info) = induc_vec_info_type; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vectorizable_induction ==="); vect_model_induction_cost (stmt_info, ncopies); @@ -5306,7 +5305,7 @@ vectorizable_induction (gimple phi, gimple_stmt_it /** Transform. **/ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "transform induction phi."); vec_def = get_initial_def_for_induction (phi); @@ -5371,7 +5370,7 @@ vectorizable_live_operation (gimple stmt, && !vect_is_simple_use (op, stmt, loop_vinfo, NULL, &def_stmt, &def, &dt)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "use not simple."); return false; @@ -5410,7 +5409,7 @@ vect_loop_kill_debug_uses (struct loop *loop, gimp { if (gimple_debug_bind_p (ustmt)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "killing debug use"); @@ -5450,7 +5449,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) bool check_profitability = false; int th; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vec_transform_loop ==="); /* Use the more conservative vectorization threshold. If the number @@ -5464,7 +5463,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) if (th >= LOOP_VINFO_VECT_FACTOR (loop_vinfo) - 1 && !LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Profitability threshold is %d loop iterations.", th); check_profitability = true; @@ -5525,7 +5524,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) for (si = gsi_start_phis (bb); !gsi_end_p (si); gsi_next (&si)) { phi = gsi_stmt (si); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "------>vectorizing phi: "); @@ -5544,12 +5543,12 @@ vect_transform_loop (loop_vec_info loop_vinfo) if ((TYPE_VECTOR_SUBPARTS (STMT_VINFO_VECTYPE (stmt_info)) != (unsigned HOST_WIDE_INT) vectorization_factor) - && dump_kind_p (MSG_NOTE)) + && dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "multiple-types."); if (STMT_VINFO_DEF_TYPE (stmt_info) == vect_induction_def) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "transform phi."); vect_transform_stmt (phi, NULL, NULL, NULL, NULL); } @@ -5565,7 +5564,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) else stmt = gsi_stmt (si); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "------>vectorizing statement: "); @@ -5637,7 +5636,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) if (!gsi_end_p (pattern_def_si)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "==> vectorizing pattern def " @@ -5664,7 +5663,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) STMT_VINFO_VECTYPE (stmt_info)); if (!STMT_SLP_TYPE (stmt_info) && nunits != (unsigned int) vectorization_factor - && dump_kind_p (MSG_NOTE)) + && dump_enabled_p ()) /* For SLP VF is set according to unrolling factor, and not to vector size, hence for SLP this print is not valid. */ dump_printf_loc (MSG_NOTE, vect_location, @@ -5678,7 +5677,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) { slp_scheduled = true; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== scheduling SLP instances ==="); @@ -5698,7 +5697,7 @@ vect_transform_loop (loop_vec_info loop_vinfo) } /* -------- vectorize statement ------------ */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "transform statement."); grouped_store = false; @@ -5741,9 +5740,9 @@ vect_transform_loop (loop_vec_info loop_vinfo) until all the loops have been transformed? */ update_ssa (TODO_update_ssa); - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "LOOP VECTORIZED."); - if (loop->inner && dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (loop->inner && dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "OUTER LOOP VECTORIZED."); } Index: tree-vect-data-refs.c =================================================================== --- tree-vect-data-refs.c (revision 192695) +++ tree-vect-data-refs.c (working copy) @@ -60,7 +60,7 @@ vect_lanes_optab_supported_p (const char *name, co if (array_mode == BLKmode) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "no array mode for %s[" HOST_WIDE_INT_PRINT_DEC "]", GET_MODE_NAME (mode), count); @@ -69,14 +69,14 @@ vect_lanes_optab_supported_p (const char *name, co if (convert_optab_handler (optab, array_mode, mode) == CODE_FOR_nothing) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "cannot use %s<%s><%s>", name, GET_MODE_NAME (array_mode), GET_MODE_NAME (mode)); return false; } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "can use %s<%s><%s>", name, GET_MODE_NAME (array_mode), GET_MODE_NAME (mode)); @@ -439,7 +439,7 @@ vect_check_interleaving (struct data_reference *dr if (diff_mod_size == 0) { vect_update_interleaving_chain (drb, dra); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Detected interleaving "); @@ -462,7 +462,7 @@ vect_check_interleaving (struct data_reference *dr if (diff_mod_size == 0) { vect_update_interleaving_chain (dra, drb); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Detected interleaving "); @@ -524,7 +524,7 @@ vect_mark_for_runtime_alias_test (ddr_p ddr, loop_ if ((unsigned) PARAM_VALUE (PARAM_VECT_MAX_VERSION_FOR_ALIAS_CHECKS) == 0) return false; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "mark for run-time aliasing test between "); @@ -535,7 +535,7 @@ vect_mark_for_runtime_alias_test (ddr_p ddr, loop_ if (optimize_loop_nest_for_size_p (loop)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "versioning not supported when optimizing for size."); return false; @@ -544,7 +544,7 @@ vect_mark_for_runtime_alias_test (ddr_p ddr, loop_ /* FORNOW: We don't support versioning with outer-loop vectorization. */ if (loop->inner) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "versioning not yet supported for outer-loops."); return false; @@ -555,7 +555,7 @@ vect_mark_for_runtime_alias_test (ddr_p ddr, loop_ if (TREE_CODE (DR_STEP (DDR_A (ddr))) != INTEGER_CST || TREE_CODE (DR_STEP (DDR_B (ddr))) != INTEGER_CST) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "versioning not yet supported for non-constant " "step"); @@ -611,7 +611,7 @@ vect_analyze_data_ref_dependence (struct data_depe if (loop_vinfo) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "versioning for alias required: " @@ -637,7 +637,7 @@ vect_analyze_data_ref_dependence (struct data_depe if (DR_IS_READ (dra) && DR_IS_READ (drb)) return false; - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "can't determine dependence between "); @@ -666,7 +666,7 @@ vect_analyze_data_ref_dependence (struct data_depe if (dra != drb && vect_check_interleaving (dra, drb)) return false; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "determined dependence between "); @@ -686,7 +686,7 @@ vect_analyze_data_ref_dependence (struct data_depe /* Loop-based vectorization and known data dependence. */ if (DDR_NUM_DIST_VECTS (ddr) == 0) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "versioning for alias required: " @@ -704,13 +704,13 @@ vect_analyze_data_ref_dependence (struct data_depe { int dist = dist_v[loop_depth]; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "dependence distance = %d.", dist); if (dist == 0) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "dependence distance == 0 between "); @@ -737,7 +737,7 @@ vect_analyze_data_ref_dependence (struct data_depe /* If DDR_REVERSED_P the order of the data-refs in DDR was reversed (to make distance vector positive), and the actual distance is negative. */ - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "dependence distance negative."); continue; @@ -749,7 +749,7 @@ vect_analyze_data_ref_dependence (struct data_depe /* The dependence distance requires reduction of the maximal vectorization factor. */ *max_vf = abs (dist); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "adjusting maximal vectorization factor to %i", *max_vf); @@ -759,13 +759,13 @@ vect_analyze_data_ref_dependence (struct data_depe { /* Dependence distance does not create dependence, as far as vectorization is concerned, in this case. */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "dependence distance >= VF."); continue; } - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized, possible dependence " @@ -795,7 +795,7 @@ vect_analyze_data_ref_dependences (loop_vec_info l VEC (ddr_p, heap) *ddrs = NULL; struct data_dependence_relation *ddr; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_analyze_dependences ==="); if (loop_vinfo) @@ -837,7 +837,7 @@ vect_compute_data_ref_alignment (struct data_refer tree misalign; tree aligned_to, alignment; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_compute_data_ref_alignment:"); @@ -870,7 +870,7 @@ vect_compute_data_ref_alignment (struct data_refer if (dr_step % GET_MODE_SIZE (TYPE_MODE (vectype)) == 0) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "inner step divides the vector-size."); misalign = STMT_VINFO_DR_INIT (stmt_info); @@ -879,7 +879,7 @@ vect_compute_data_ref_alignment (struct data_refer } else { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "inner step doesn't divide the vector-size."); misalign = NULL_TREE; @@ -898,7 +898,7 @@ vect_compute_data_ref_alignment (struct data_refer if (dr_step % GET_MODE_SIZE (TYPE_MODE (vectype)) != 0) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "SLP: step doesn't divide the vector-size."); misalign = NULL_TREE; @@ -911,7 +911,7 @@ vect_compute_data_ref_alignment (struct data_refer if ((aligned_to && tree_int_cst_compare (aligned_to, alignment) < 0) || !misalign) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Unknown alignment for access: "); @@ -941,7 +941,7 @@ vect_compute_data_ref_alignment (struct data_refer if (!vect_can_force_dr_alignment_p (base, TYPE_ALIGN (vectype)) || (TREE_STATIC (base) && flag_section_anchors)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "can't force alignment of ref: "); @@ -953,7 +953,7 @@ vect_compute_data_ref_alignment (struct data_refer /* Force the alignment of the decl. NOTE: This is the only change to the code we make during the analysis phase, before deciding to vectorize the loop. */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "force alignment of "); dump_generic_expr (MSG_NOTE, TDF_SLIM, ref); @@ -987,7 +987,7 @@ vect_compute_data_ref_alignment (struct data_refer if (!host_integerp (misalign, 1)) { /* Negative or overflowed misalignment value. */ - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "unexpected misalign value"); return false; @@ -995,7 +995,7 @@ vect_compute_data_ref_alignment (struct data_refer SET_DR_MISALIGNMENT (dr, TREE_INT_CST_LOW (misalign)); - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "misalign = %d bytes of ref ", DR_MISALIGNMENT (dr)); @@ -1095,7 +1095,7 @@ vect_update_misalignment_for_peel (struct data_ref return; } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Setting misalignment to -1."); SET_DR_MISALIGNMENT (dr, -1); } @@ -1142,7 +1142,7 @@ vect_verify_datarefs_alignment (loop_vec_info loop supportable_dr_alignment = vect_supportable_dr_alignment (dr, false); if (!supportable_dr_alignment) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { if (DR_IS_READ (dr)) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, @@ -1157,8 +1157,7 @@ vect_verify_datarefs_alignment (loop_vec_info loop } return false; } - if (supportable_dr_alignment != dr_aligned - && dump_kind_p (MSG_NOTE)) + if (supportable_dr_alignment != dr_aligned && dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Vectorizing an unaligned access."); } @@ -1215,7 +1214,7 @@ vector_alignment_reachable_p (struct data_referenc { HOST_WIDE_INT elmsize = int_cst_value (TYPE_SIZE_UNIT (TREE_TYPE (vectype))); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "data size =" HOST_WIDE_INT_PRINT_DEC, elmsize); @@ -1224,7 +1223,7 @@ vector_alignment_reachable_p (struct data_referenc } if (DR_MISALIGNMENT (dr) % elmsize) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "data size does not divide the misalignment.\n"); return false; @@ -1235,7 +1234,7 @@ vector_alignment_reachable_p (struct data_referenc { tree type = TREE_TYPE (DR_REF (dr)); bool is_packed = not_size_aligned (DR_REF (dr)); - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Unknown misalignment, is_packed = %d",is_packed); if (targetm.vectorize.vector_alignment_reachable (type, is_packed)) @@ -1269,7 +1268,7 @@ vect_get_data_access_cost (struct data_reference * else vect_get_store_cost (dr, ncopies, inside_cost, body_cost_vec); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_get_data_access_cost: inside_cost = %d, " "outside_cost = %d.", *inside_cost, *outside_cost); @@ -1567,7 +1566,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo unsigned int nelements, mis, same_align_drs_max = 0; stmt_vector_for_cost body_cost_vec = NULL; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_enhance_data_refs_alignment ==="); @@ -1622,7 +1621,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo and so we can't generate the new base for the pointer. */ if (STMT_VINFO_STRIDE_LOAD_P (stmt_info)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "strided load prevents peeling"); do_peeling = false; @@ -1738,7 +1737,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo { if (!aligned_access_p (dr)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "vector alignment may not be reachable"); break; @@ -1879,7 +1878,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo if (STMT_VINFO_GROUPED_ACCESS (stmt_info)) npeel /= GROUP_SIZE (stmt_info); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Try peeling by %d", npeel); } @@ -1951,7 +1950,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo else LOOP_PEELING_FOR_ALIGNMENT (loop_vinfo) = DR_MISALIGNMENT (dr0); SET_DR_MISALIGNMENT (dr0, 0); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Alignment of access forced using peeling."); @@ -2077,12 +2076,12 @@ vect_enhance_data_refs_alignment (loop_vec_info lo stmt_vec_info stmt_info = vinfo_for_stmt (stmt); dr = STMT_VINFO_DATA_REF (stmt_info); SET_DR_MISALIGNMENT (dr, 0); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Alignment of access forced using versioning."); } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Versioning for alignment will be applied."); @@ -2148,7 +2147,7 @@ vect_find_same_alignment_drs (struct data_dependen { int dist = dist_v[loop_depth]; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "dependence distance = %d.", dist); @@ -2159,7 +2158,7 @@ vect_find_same_alignment_drs (struct data_dependen /* Two references with distance zero have the same alignment. */ VEC_safe_push (dr_p, heap, STMT_VINFO_SAME_ALIGN_REFS (stmtinfo_a), drb); VEC_safe_push (dr_p, heap, STMT_VINFO_SAME_ALIGN_REFS (stmtinfo_b), dra); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "accesses have the same alignment."); @@ -2183,7 +2182,7 @@ bool vect_analyze_data_refs_alignment (loop_vec_info loop_vinfo, bb_vec_info bb_vinfo) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_analyze_data_refs_alignment ==="); @@ -2201,7 +2200,7 @@ vect_analyze_data_refs_alignment (loop_vec_info lo if (!vect_compute_data_refs_alignment (loop_vinfo, bb_vinfo)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: can't calculate alignment " "for data ref."); @@ -2254,7 +2253,7 @@ vect_analyze_group_access (struct data_reference * { GROUP_FIRST_ELEMENT (vinfo_for_stmt (stmt)) = stmt; GROUP_SIZE (vinfo_for_stmt (stmt)) = groupsize; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Detected single element interleaving "); @@ -2265,13 +2264,13 @@ vect_analyze_group_access (struct data_reference * if (loop_vinfo) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Data access with gaps requires scalar " "epilogue loop"); if (loop->inner) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Peeling for outer loop is not" " supported"); @@ -2284,7 +2283,7 @@ vect_analyze_group_access (struct data_reference * return true; } - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not consecutive access "); @@ -2324,7 +2323,7 @@ vect_analyze_group_access (struct data_reference * { if (DR_IS_WRITE (data_ref)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Two store stmts share the same dr."); return false; @@ -2335,7 +2334,7 @@ vect_analyze_group_access (struct data_reference * if (GROUP_READ_WRITE_DEPENDENCE (vinfo_for_stmt (next)) || GROUP_READ_WRITE_DEPENDENCE (vinfo_for_stmt (prev))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "READ_WRITE dependence in interleaving."); return false; @@ -2355,7 +2354,7 @@ vect_analyze_group_access (struct data_reference * next_step = DR_STEP (STMT_VINFO_DATA_REF (vinfo_for_stmt (next))); if (tree_int_cst_compare (step, next_step)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not consecutive access in interleaving"); return false; @@ -2372,7 +2371,7 @@ vect_analyze_group_access (struct data_reference * slp_impossible = true; if (DR_IS_WRITE (data_ref)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "interleaved store with gaps"); return false; @@ -2401,7 +2400,7 @@ vect_analyze_group_access (struct data_reference * greater than STEP. */ if (dr_step && dr_step < count_in_bytes + gaps * type_size) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "interleaving size is greater than step for "); @@ -2424,7 +2423,7 @@ vect_analyze_group_access (struct data_reference * } else { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "interleaved store with gaps"); return false; @@ -2434,7 +2433,7 @@ vect_analyze_group_access (struct data_reference * /* Check that STEP is a multiple of type size. */ if (dr_step && (dr_step % type_size) != 0) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "step is not a multiple of type size: step "); @@ -2450,7 +2449,7 @@ vect_analyze_group_access (struct data_reference * groupsize = count; GROUP_SIZE (vinfo_for_stmt (stmt)) = groupsize; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Detected interleaving of size %d", (int)groupsize); @@ -2469,13 +2468,13 @@ vect_analyze_group_access (struct data_reference * /* There is a gap in the end of the group. */ if (groupsize - last_accessed_element > 0 && loop_vinfo) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Data access with gaps requires scalar " "epilogue loop"); if (loop->inner) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Peeling for outer loop is not supported"); return false; @@ -2508,7 +2507,7 @@ vect_analyze_data_ref_access (struct data_referenc if (loop_vinfo && !step) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "bad data-ref access in loop"); return false; @@ -2531,7 +2530,7 @@ vect_analyze_data_ref_access (struct data_referenc step = STMT_VINFO_DR_STEP (stmt_info); if (integer_zerop (step)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "zero step in outer loop."); if (DR_IS_READ (dr)) @@ -2557,7 +2556,7 @@ vect_analyze_data_ref_access (struct data_referenc if (loop && nested_in_vect_loop_p (loop, stmt)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "grouped access in outer loop."); return false; @@ -2588,7 +2587,7 @@ vect_analyze_data_ref_accesses (loop_vec_info loop VEC (data_reference_p, heap) *datarefs; struct data_reference *dr; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_analyze_data_ref_accesses ==="); @@ -2601,7 +2600,7 @@ vect_analyze_data_ref_accesses (loop_vec_info loop if (STMT_VINFO_VECTORIZABLE (vinfo_for_stmt (DR_STMT (dr))) && !vect_analyze_data_ref_access (dr)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: complicated access pattern."); @@ -2631,7 +2630,7 @@ vect_prune_runtime_alias_test_list (loop_vec_info LOOP_VINFO_MAY_ALIAS_DDRS (loop_vinfo); unsigned i, j; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_prune_runtime_alias_test_list ==="); @@ -2649,7 +2648,7 @@ vect_prune_runtime_alias_test_list (loop_vec_info if (vect_vfa_range_equal (ddr_i, ddr_j)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "found equal ranges "); @@ -2677,7 +2676,7 @@ vect_prune_runtime_alias_test_list (loop_vec_info if (VEC_length (ddr_p, ddrs) > (unsigned) PARAM_VALUE (PARAM_VECT_MAX_VERSION_FOR_ALIAS_CHECKS)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "disable versioning for alias - max number of " @@ -2964,7 +2963,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, tree scalar_type; bool res, stop_bb_analysis = false; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_analyze_data_refs ===\n"); @@ -2979,7 +2978,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (!res) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: loop contains function calls" " or data references that cannot be analyzed"); @@ -3011,7 +3010,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (!compute_all_dependences (BB_VINFO_DATAREFS (bb_vinfo), &BB_VINFO_DDRS (bb_vinfo), NULL, true)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: basic block contains function" " calls or data references that cannot be" @@ -3035,7 +3034,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (!dr || !DR_REF (dr)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: unhandled data-ref "); return false; @@ -3081,7 +3080,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (!gather) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: data ref analysis " @@ -3102,7 +3101,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (TREE_CODE (DR_BASE_ADDRESS (dr)) == INTEGER_CST) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: base addr of dr is a " "constant"); @@ -3121,7 +3120,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (TREE_THIS_VOLATILE (DR_REF (dr))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: volatile type "); @@ -3140,7 +3139,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (stmt_can_throw_internal (stmt)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: statement can throw an " @@ -3163,7 +3162,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (TREE_CODE (DR_REF (dr)) == COMPONENT_REF && DECL_BIT_FIELD (TREE_OPERAND (DR_REF (dr), 1))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: statement is bitfield " @@ -3189,7 +3188,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (is_gimple_call (stmt)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: dr in a call "); @@ -3232,7 +3231,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, tree inner_base = build_fold_indirect_ref (fold_build_pointer_plus (base, init)); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "analyze in outer-loop: "); @@ -3245,7 +3244,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (pbitpos % BITS_PER_UNIT != 0) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "failed: bit offset alignment.\n"); return false; @@ -3255,7 +3254,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (!simple_iv (loop, loop_containing_stmt (stmt), outer_base, &base_iv, false)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "failed: evolution of base is not affine.\n"); return false; @@ -3278,7 +3277,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, else if (!simple_iv (loop, loop_containing_stmt (stmt), poffset, &offset_iv, false)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "evolution of offset is not affine.\n"); return false; @@ -3303,7 +3302,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, STMT_VINFO_DR_ALIGNED_TO (stmt_info) = size_int (highest_pow2_factor (offset_iv.base)); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "\touter base_address: "); @@ -3327,7 +3326,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (STMT_VINFO_DATA_REF (stmt_info)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: more than one data ref " @@ -3355,7 +3354,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, get_vectype_for_scalar_type (scalar_type); if (!STMT_VINFO_VECTYPE (stmt_info)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: no vectype for stmt: "); @@ -3406,7 +3405,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, { STMT_VINFO_DATA_REF (stmt_info) = NULL; free_data_ref (dr); - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: not suitable for gather " @@ -3459,7 +3458,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, if (bad) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: data dependence conflict" @@ -3480,7 +3479,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo, = vect_check_strided_load (stmt, loop_vinfo, NULL, NULL); if (!strided_load) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: not suitable for strided " @@ -3668,7 +3667,7 @@ vect_create_addr_base_for_vector_ref (gimple stmt, mark_ptr_info_alignment_unknown (SSA_NAME_PTR_INFO (vec_stmt)); } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "created "); dump_generic_expr (MSG_NOTE, TDF_SLIM, vec_stmt); @@ -3790,7 +3789,7 @@ vect_create_data_ref_ptr (gimple stmt, tree aggr_t in LOOP. */ base_name = build_fold_indirect_ref (unshare_expr (DR_BASE_ADDRESS (dr))); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { tree data_ref_base = base_name; dump_printf_loc (MSG_NOTE, vect_location, @@ -4120,7 +4119,7 @@ vect_grouped_store_supported (tree vectype, unsign /* vect_permute_store_chain requires the group size to be a power of two. */ if (exact_log2 (count) == -1) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "the size of the group of accesses" " is not a power of 2"); @@ -4146,7 +4145,7 @@ vect_grouped_store_supported (tree vectype, unsign } } - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf (MSG_MISSED_OPTIMIZATION, "interleave op not supported by target."); return false; @@ -4564,7 +4563,7 @@ vect_grouped_load_supported (tree vectype, unsigne /* vect_permute_load_chain requires the group size to be a power of two. */ if (exact_log2 (count) == -1) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "the size of the group of accesses" " is not a power of 2"); @@ -4588,7 +4587,7 @@ vect_grouped_load_supported (tree vectype, unsigne } } - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "extract even/odd not supported by target"); return false; Index: tree-vect-patterns.c =================================================================== --- tree-vect-patterns.c (revision 192695) +++ tree-vect-patterns.c (working copy) @@ -416,7 +416,7 @@ vect_recog_dot_prod_pattern (VEC (gimple, heap) ** pattern_stmt = gimple_build_assign_with_ops (DOT_PROD_EXPR, var, oprnd00, oprnd01, oprnd1); - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "vect_recog_dot_prod_pattern: detected: "); @@ -676,7 +676,7 @@ vect_recog_widen_mult_pattern (VEC (gimple, heap) return NULL; /* Pattern detected. */ - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "vect_recog_widen_mult_pattern: detected: "); @@ -699,7 +699,7 @@ vect_recog_widen_mult_pattern (VEC (gimple, heap) pattern_stmt = gimple_build_assign_with_ops (WIDEN_MULT_EXPR, var, oprnd0, oprnd1); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_gimple_stmt_loc (MSG_NOTE, vect_location, TDF_SLIM, pattern_stmt, 0); VEC_safe_push (gimple, heap, *stmts, last_stmt); @@ -912,7 +912,7 @@ vect_recog_widen_sum_pattern (VEC (gimple, heap) * pattern_stmt = gimple_build_assign_with_ops (WIDEN_SUM_EXPR, var, oprnd0, oprnd1); - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "vect_recog_widen_sum_pattern: detected: "); @@ -1217,7 +1217,7 @@ vect_recog_over_widening_pattern (VEC (gimple, hea STMT_VINFO_RELATED_STMT (vinfo_for_stmt (stmt)) = pattern_stmt; new_pattern_def_seq (vinfo_for_stmt (stmt), new_def_stmt); - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "created pattern stmt: "); @@ -1285,7 +1285,7 @@ vect_recog_over_widening_pattern (VEC (gimple, hea return NULL; /* Pattern detected. */ - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "vect_recog_over_widening_pattern: detected: "); @@ -1421,7 +1421,7 @@ vect_recog_widen_shift_pattern (VEC (gimple, heap) return NULL; /* Pattern detected. */ - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "vect_recog_widen_shift_pattern: detected: "); @@ -1445,7 +1445,7 @@ vect_recog_widen_shift_pattern (VEC (gimple, heap) pattern_stmt = gimple_build_assign_with_ops (WIDEN_LSHIFT_EXPR, var, oprnd0, oprnd1); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_gimple_stmt_loc (MSG_NOTE, vect_location, TDF_SLIM, pattern_stmt, 0); VEC_safe_push (gimple, heap, *stmts, last_stmt); @@ -1567,7 +1567,7 @@ vect_recog_vector_vector_shift_pattern (VEC (gimpl } /* Pattern detected. */ - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "vect_recog_vector_vector_shift_pattern: detected: "); @@ -1575,7 +1575,7 @@ vect_recog_vector_vector_shift_pattern (VEC (gimpl var = vect_recog_temp_ssa_var (TREE_TYPE (oprnd0), NULL); pattern_stmt = gimple_build_assign_with_ops (rhs_code, var, oprnd0, def); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_gimple_stmt_loc (MSG_NOTE, vect_location, TDF_SLIM, pattern_stmt, 0); VEC_safe_push (gimple, heap, *stmts, last_stmt); @@ -1685,7 +1685,7 @@ vect_recog_divmod_pattern (VEC (gimple, heap) **st return NULL; /* Pattern detected. */ - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "vect_recog_divmod_pattern: detected: "); @@ -1789,7 +1789,7 @@ vect_recog_divmod_pattern (VEC (gimple, heap) **st signmask); } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_gimple_stmt_loc (MSG_NOTE, vect_location, TDF_SLIM, pattern_stmt, 0); @@ -2031,7 +2031,7 @@ vect_recog_divmod_pattern (VEC (gimple, heap) **st } /* Pattern detected. */ - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "vect_recog_divmod_pattern: detected: "); @@ -2199,7 +2199,7 @@ vect_recog_mixed_size_cond_pattern (VEC (gimple, h *type_in = vecitype; *type_out = vectype; - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "vect_recog_mixed_size_cond_pattern: detected: "); @@ -2592,7 +2592,7 @@ vect_recog_bool_pattern (VEC (gimple, heap) **stmt *type_out = vectype; *type_in = vectype; VEC_safe_push (gimple, heap, *stmts, last_stmt); - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "vect_recog_bool_pattern: detected: "); @@ -2638,7 +2638,7 @@ vect_recog_bool_pattern (VEC (gimple, heap) **stmt *type_out = vectype; *type_in = vectype; VEC_safe_push (gimple, heap, *stmts, last_stmt); - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "vect_recog_bool_pattern: detected: "); return pattern_stmt; @@ -2788,7 +2788,7 @@ vect_pattern_recog_1 (vect_recog_func_ptr vect_rec } /* Found a vectorizable pattern. */ - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "pattern recognized: "); @@ -2814,7 +2814,7 @@ vect_pattern_recog_1 (vect_recog_func_ptr vect_rec { stmt_info = vinfo_for_stmt (stmt); pattern_stmt = STMT_VINFO_RELATED_STMT (stmt_info); - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "additional pattern stmt: "); @@ -2915,7 +2915,7 @@ vect_pattern_recog (loop_vec_info loop_vinfo, bb_v VEC (gimple, heap) *stmts_to_replace = VEC_alloc (gimple, heap, 1); gimple stmt; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_pattern_recog ==="); Index: tree-vect-stmts.c =================================================================== --- tree-vect-stmts.c (revision 192695) +++ tree-vect-stmts.c (working copy) @@ -190,7 +190,7 @@ vect_mark_relevant (VEC(gimple,heap) **worklist, g bool save_live_p = STMT_VINFO_LIVE_P (stmt_info); gimple pattern_stmt; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "mark relevant %d, live %d.", relevant, live_p); @@ -246,7 +246,7 @@ vect_mark_relevant (VEC(gimple,heap) **worklist, g pattern_stmt = STMT_VINFO_RELATED_STMT (stmt_info); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "last stmt in pattern. don't mark" " relevant/live."); @@ -265,7 +265,7 @@ vect_mark_relevant (VEC(gimple,heap) **worklist, g if (STMT_VINFO_RELEVANT (stmt_info) == save_relevant && STMT_VINFO_LIVE_P (stmt_info) == save_live_p) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "already marked relevant/live."); return; @@ -310,7 +310,7 @@ vect_stmt_relevant_p (gimple stmt, loop_vec_info l if (gimple_code (stmt) != GIMPLE_PHI) if (gimple_vdef (stmt)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vec_stmt_relevant_p: stmt has vdefs."); *relevant = vect_used_in_scope; @@ -324,7 +324,7 @@ vect_stmt_relevant_p (gimple stmt, loop_vec_info l basic_block bb = gimple_bb (USE_STMT (use_p)); if (!flow_bb_inside_loop_p (loop, bb)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vec_stmt_relevant_p: used out of loop."); @@ -437,7 +437,7 @@ process_use (gimple stmt, tree use, loop_vec_info if (!vect_is_simple_use (use, stmt, loop_vinfo, NULL, &def_stmt, &def, &dt)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: unsupported use in stmt."); return false; @@ -449,7 +449,7 @@ process_use (gimple stmt, tree use, loop_vec_info def_bb = gimple_bb (def_stmt); if (!flow_bb_inside_loop_p (loop, def_bb)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "def_stmt is out of loop."); return true; } @@ -467,7 +467,7 @@ process_use (gimple stmt, tree use, loop_vec_info && STMT_VINFO_DEF_TYPE (dstmt_vinfo) == vect_reduction_def && bb->loop_father == def_bb->loop_father) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "reduc-stmt defining reduc-phi in the same nest."); if (STMT_VINFO_IN_PATTERN_P (dstmt_vinfo)) @@ -487,7 +487,7 @@ process_use (gimple stmt, tree use, loop_vec_info ... */ if (flow_loop_nested_p (def_bb->loop_father, bb->loop_father)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "outer-loop def-stmt defining inner-loop stmt."); @@ -525,7 +525,7 @@ process_use (gimple stmt, tree use, loop_vec_info stmt # use (d) */ else if (flow_loop_nested_p (bb->loop_father, def_bb->loop_father)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "inner-loop def-stmt defining outer-loop stmt."); @@ -589,7 +589,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo enum vect_relevant relevant, tmp_relevant; enum vect_def_type def_type; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_mark_stmts_to_be_vectorized ==="); @@ -602,7 +602,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo for (si = gsi_start_phis (bb); !gsi_end_p (si); gsi_next (&si)) { phi = gsi_stmt (si); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "init: phi relevant? "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0); @@ -614,7 +614,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo for (si = gsi_start_bb (bb); !gsi_end_p (si); gsi_next (&si)) { stmt = gsi_stmt (si); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "init: stmt relevant? "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, stmt, 0); @@ -632,7 +632,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo ssa_op_iter iter; stmt = VEC_pop (gimple, worklist); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "worklist: examine stmt: "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, stmt, 0); @@ -677,7 +677,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo /* fall through */ default: - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "unsupported use of reduction."); VEC_free (gimple, heap, worklist); @@ -692,7 +692,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo && tmp_relevant != vect_used_in_outer_by_reduction && tmp_relevant != vect_used_in_outer) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "unsupported use of nested cycle."); @@ -707,7 +707,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo if (tmp_relevant != vect_unused_in_scope && tmp_relevant != vect_used_by_reduction) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "unsupported use of double reduction."); @@ -830,7 +830,7 @@ vect_model_simple_cost (stmt_vec_info stmt_info, i inside_cost = record_stmt_cost (body_cost_vec, ncopies, vector_stmt, stmt_info, 0, vect_body); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_simple_cost: inside_cost = %d, " "prologue_cost = %d .", inside_cost, prologue_cost); @@ -876,7 +876,7 @@ vect_model_promotion_demotion_cost (stmt_vec_info prologue_cost += add_stmt_cost (target_cost_data, 1, vector_stmt, stmt_info, 0, vect_prologue); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_promotion_demotion_cost: inside_cost = %d, " "prologue_cost = %d .", inside_cost, prologue_cost); @@ -960,7 +960,7 @@ vect_model_store_cost (stmt_vec_info stmt_info, in inside_cost = record_stmt_cost (body_cost_vec, nstmts, vec_perm, stmt_info, 0, vect_body); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_store_cost: strided group_size = %d .", group_size); @@ -969,7 +969,7 @@ vect_model_store_cost (stmt_vec_info stmt_info, in /* Costs of the stores. */ vect_get_store_cost (first_dr, ncopies, &inside_cost, body_cost_vec); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_store_cost: inside_cost = %d, " "prologue_cost = %d .", inside_cost, prologue_cost); @@ -994,7 +994,7 @@ vect_get_store_cost (struct data_reference *dr, in vector_store, stmt_info, 0, vect_body); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_store_cost: aligned."); break; @@ -1006,7 +1006,7 @@ vect_get_store_cost (struct data_reference *dr, in *inside_cost += record_stmt_cost (body_cost_vec, ncopies, unaligned_store, stmt_info, DR_MISALIGNMENT (dr), vect_body); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_store_cost: unaligned supported by " "hardware."); @@ -1017,7 +1017,7 @@ vect_get_store_cost (struct data_reference *dr, in { *inside_cost = VECT_MAX_COST; - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "vect_model_store_cost: unsupported access."); break; @@ -1076,7 +1076,7 @@ vect_model_load_cost (stmt_vec_info stmt_info, int inside_cost += record_stmt_cost (body_cost_vec, nstmts, vec_perm, stmt_info, 0, vect_body); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_load_cost: strided group_size = %d .", group_size); @@ -1100,7 +1100,7 @@ vect_model_load_cost (stmt_vec_info stmt_info, int &inside_cost, &prologue_cost, prologue_cost_vec, body_cost_vec, true); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_load_cost: inside_cost = %d, " "prologue_cost = %d .", inside_cost, prologue_cost); @@ -1127,7 +1127,7 @@ vect_get_load_cost (struct data_reference *dr, int *inside_cost += record_stmt_cost (body_cost_vec, ncopies, vector_load, stmt_info, 0, vect_body); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_load_cost: aligned."); @@ -1140,7 +1140,7 @@ vect_get_load_cost (struct data_reference *dr, int unaligned_load, stmt_info, DR_MISALIGNMENT (dr), vect_body); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_load_cost: unaligned supported by " "hardware."); @@ -1161,7 +1161,7 @@ vect_get_load_cost (struct data_reference *dr, int *inside_cost += record_stmt_cost (body_cost_vec, 1, vector_stmt, stmt_info, 0, vect_body); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_load_cost: explicit realign"); @@ -1169,7 +1169,7 @@ vect_get_load_cost (struct data_reference *dr, int } case dr_explicit_realign_optimized: { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_load_cost: unaligned software " "pipelined."); @@ -1197,7 +1197,7 @@ vect_get_load_cost (struct data_reference *dr, int *inside_cost += record_stmt_cost (body_cost_vec, ncopies, vec_perm, stmt_info, 0, vect_body); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vect_model_load_cost: explicit realign optimized"); @@ -1208,7 +1208,7 @@ vect_get_load_cost (struct data_reference *dr, int { *inside_cost = VECT_MAX_COST; - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "vect_model_load_cost: unsupported access."); break; @@ -1258,7 +1258,7 @@ vect_init_vector_1 (gimple stmt, gimple new_stmt, } } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "created new init_stmt: "); @@ -1340,7 +1340,7 @@ vect_get_vec_def_for_operand (tree op, gimple stmt bool is_simple_use; tree vector_type; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "vect_get_vec_def_for_operand: "); @@ -1350,7 +1350,7 @@ vect_get_vec_def_for_operand (tree op, gimple stmt is_simple_use = vect_is_simple_use (op, stmt, loop_vinfo, NULL, &def_stmt, &def, &dt); gcc_assert (is_simple_use); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { int loc_printed = 0; if (def) @@ -1382,7 +1382,7 @@ vect_get_vec_def_for_operand (tree op, gimple stmt *scalar_def = op; /* Create 'vect_cst_ = {cst,cst,...,cst}' */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Create vector_cst. nunits = %d", nunits); @@ -1399,7 +1399,7 @@ vect_get_vec_def_for_operand (tree op, gimple stmt *scalar_def = def; /* Create 'vec_inv = {inv,inv,..,inv}' */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Create vector_inv."); return vect_init_vector (stmt, def, vector_type, NULL); @@ -1661,7 +1661,7 @@ vect_finish_stmt_generation (gimple stmt, gimple v set_vinfo_for_stmt (vec_stmt, new_stmt_vec_info (vec_stmt, loop_vinfo, bb_vinfo)); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "add new stmt: "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, vec_stmt, 0); @@ -1764,7 +1764,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat if (rhs_type && !types_compatible_p (rhs_type, TREE_TYPE (op))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "argument types differ."); return false; @@ -1775,7 +1775,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat if (!vect_is_simple_use_1 (op, stmt, loop_vinfo, bb_vinfo, &def_stmt, &def, &dt[i], &opvectype)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "use not simple."); return false; @@ -1786,7 +1786,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat else if (opvectype && opvectype != vectype_in) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "argument vector types differ."); return false; @@ -1800,7 +1800,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat gcc_assert (vectype_in); if (!vectype_in) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "no vectype for scalar type "); @@ -1829,7 +1829,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat fndecl = vectorizable_function (stmt, vectype_out, vectype_in); if (fndecl == NULL_TREE) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "function is not vectorizable."); @@ -1852,7 +1852,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat if (!vec_stmt) /* transformation not required. */ { STMT_VINFO_TYPE (stmt_info) = call_vec_info_type; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vectorizable_call ==="); vect_model_simple_cost (stmt_info, ncopies, dt, NULL, NULL); return true; @@ -1860,7 +1860,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat /** Transform. **/ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "transform call."); /* Handle def. */ @@ -2375,7 +2375,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ && (TYPE_PRECISION (rhs_type) != GET_MODE_PRECISION (TYPE_MODE (rhs_type))))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "type conversion to/from bit-precision unsupported."); return false; @@ -2385,7 +2385,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ if (!vect_is_simple_use_1 (op0, stmt, loop_vinfo, bb_vinfo, &def_stmt, &def, &dt[0], &vectype_in)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "use not simple."); return false; @@ -2407,7 +2407,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ if (!ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "use not simple."); return false; @@ -2422,7 +2422,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ gcc_assert (vectype_in); if (!vectype_in) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "no vectype for scalar type "); @@ -2466,7 +2466,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ break; /* FALLTHRU */ unsupported: - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "conversion not supported by target."); return false; @@ -2565,7 +2565,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ if (!vec_stmt) /* transformation not required. */ { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vectorizable_conversion ==="); if (code == FIX_TRUNC_EXPR || code == FLOAT_EXPR) @@ -2588,7 +2588,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_ } /** Transform. **/ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "transform conversion. ncopies = %d.", ncopies); @@ -2941,7 +2941,7 @@ vectorizable_assignment (gimple stmt, gimple_stmt_ if (!vect_is_simple_use_1 (op, stmt, loop_vinfo, bb_vinfo, &def_stmt, &def, &dt[0], &vectype_in)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "use not simple."); return false; @@ -2970,7 +2970,7 @@ vectorizable_assignment (gimple stmt, gimple_stmt_ > TYPE_PRECISION (TREE_TYPE (op))) && TYPE_UNSIGNED (TREE_TYPE (op)))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "type conversion to/from bit-precision " "unsupported."); @@ -2980,7 +2980,7 @@ vectorizable_assignment (gimple stmt, gimple_stmt_ if (!vec_stmt) /* transformation not required. */ { STMT_VINFO_TYPE (stmt_info) = assignment_vec_info_type; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vectorizable_assignment ==="); vect_model_simple_cost (stmt_info, ncopies, dt, NULL, NULL); @@ -2988,7 +2988,7 @@ vectorizable_assignment (gimple stmt, gimple_stmt_ } /** Transform. **/ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "transform assignment."); /* Handle def. */ @@ -3135,7 +3135,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera if (TYPE_PRECISION (TREE_TYPE (scalar_dest)) != GET_MODE_PRECISION (TYPE_MODE (TREE_TYPE (scalar_dest)))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "bit-precision shifts not supported."); return false; @@ -3145,7 +3145,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera if (!vect_is_simple_use_1 (op0, stmt, loop_vinfo, bb_vinfo, &def_stmt, &def, &dt[0], &vectype)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "use not simple."); return false; @@ -3158,7 +3158,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera gcc_assert (vectype); if (!vectype) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "no vectype for scalar type "); return false; @@ -3173,7 +3173,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera if (!vect_is_simple_use_1 (op1, stmt, loop_vinfo, bb_vinfo, &def_stmt, &def, &dt[1], &op1_vectype)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "use not simple."); return false; @@ -3218,7 +3218,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera } else { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "operand mode requires invariant argument."); return false; @@ -3228,7 +3228,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera if (!scalar_shift_arg) { optab = optab_for_tree_code (code, vectype, optab_vector); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vector/vector shift/rotate found."); @@ -3237,7 +3237,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera if (op1_vectype == NULL_TREE || TYPE_MODE (op1_vectype) != TYPE_MODE (vectype)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "unusable type for last operand in" " vector/vector shift/rotate."); @@ -3252,7 +3252,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera if (optab && optab_handler (optab, TYPE_MODE (vectype)) != CODE_FOR_nothing) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vector/scalar shift/rotate found."); } @@ -3265,7 +3265,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera { scalar_shift_arg = false; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "vector/vector shift/rotate found."); @@ -3282,7 +3282,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera && TYPE_MODE (TREE_TYPE (vectype)) != TYPE_MODE (TREE_TYPE (op1))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "unusable type for last operand in" " vector/vector shift/rotate."); @@ -3302,7 +3302,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera /* Supportable by target? */ if (!optab) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "no optab."); return false; @@ -3311,7 +3311,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera icode = (int) optab_handler (optab, vec_mode); if (icode == CODE_FOR_nothing) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "op not supported by target."); /* Check only during analysis. */ @@ -3319,7 +3319,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera || (vf < vect_min_worthwhile_factor (code) && !vec_stmt)) return false; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "proceeding using word mode."); } @@ -3328,7 +3328,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera && vf < vect_min_worthwhile_factor (code) && !vec_stmt) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not worthwhile without SIMD support."); return false; @@ -3337,7 +3337,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera if (!vec_stmt) /* transformation not required. */ { STMT_VINFO_TYPE (stmt_info) = shift_vec_info_type; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vectorizable_shift ==="); vect_model_simple_cost (stmt_info, ncopies, dt, NULL, NULL); return true; @@ -3345,7 +3345,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera /** Transform. **/ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "transform binary/unary operation."); @@ -3382,7 +3382,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera optab_op2_mode = insn_data[icode].operand[2].mode; if (!VECTOR_MODE_P (optab_op2_mode)) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "operand 1 using scalar mode."); vec_oprnd1 = op1; @@ -3510,7 +3510,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i op_type = TREE_CODE_LENGTH (code); if (op_type != unary_op && op_type != binary_op && op_type != ternary_op) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "num. args = %d (not unary/binary/ternary op).", op_type); @@ -3529,7 +3529,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i && code != BIT_XOR_EXPR && code != BIT_AND_EXPR) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "bit-precision arithmetic not supported."); return false; @@ -3539,7 +3539,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i if (!vect_is_simple_use_1 (op0, stmt, loop_vinfo, bb_vinfo, &def_stmt, &def, &dt[0], &vectype)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "use not simple."); return false; @@ -3552,7 +3552,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i gcc_assert (vectype); if (!vectype) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "no vectype for scalar type "); @@ -3574,7 +3574,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i if (!vect_is_simple_use (op1, stmt, loop_vinfo, bb_vinfo, &def_stmt, &def, &dt[1])) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "use not simple."); return false; @@ -3586,7 +3586,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i if (!vect_is_simple_use (op2, stmt, loop_vinfo, bb_vinfo, &def_stmt, &def, &dt[2])) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "use not simple."); return false; @@ -3628,7 +3628,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i optab = optab_for_tree_code (code, vectype, optab_default); if (!optab) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "no optab."); return false; @@ -3638,14 +3638,14 @@ vectorizable_operation (gimple stmt, gimple_stmt_i if (icode == CODE_FOR_nothing) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "op not supported by target."); /* Check only during analysis. */ if (GET_MODE_SIZE (vec_mode) != UNITS_PER_WORD || (!vec_stmt && vf < vect_min_worthwhile_factor (code))) return false; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "proceeding using word mode."); } @@ -3654,7 +3654,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i && !vec_stmt && vf < vect_min_worthwhile_factor (code)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not worthwhile without SIMD support."); return false; @@ -3663,7 +3663,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i if (!vec_stmt) /* transformation not required. */ { STMT_VINFO_TYPE (stmt_info) = op_vec_info_type; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vectorizable_operation ==="); vect_model_simple_cost (stmt_info, ncopies, dt, NULL, NULL); @@ -3672,7 +3672,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i /** Transform. **/ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "transform binary/unary operation."); @@ -3860,7 +3860,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera /* FORNOW. This restriction should be relaxed. */ if (loop && nested_in_vect_loop_p (loop, stmt) && ncopies > 1) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "multiple types in nested loop."); return false; @@ -3894,7 +3894,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera if (!vect_is_simple_use (op, stmt, loop_vinfo, bb_vinfo, &def_stmt, &def, &dt)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "use not simple."); return false; @@ -3915,7 +3915,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera ? STMT_VINFO_DR_STEP (stmt_info) : DR_STEP (dr), size_zero_node) < 0) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "negative step for store."); return false; @@ -3946,7 +3946,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera if (!vect_is_simple_use (op, next_stmt, loop_vinfo, bb_vinfo, &def_stmt, &def, &dt)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "use not simple."); return false; @@ -4008,7 +4008,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera group_size = vec_num = 1; } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "transform store. ncopies = %d", ncopies); @@ -4396,7 +4396,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat /* FORNOW. This restriction should be relaxed. */ if (nested_in_vect_loop && ncopies > 1) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "multiple types in nested loop."); return false; @@ -4436,7 +4436,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat (e.g. - data copies). */ if (optab_handler (mov_optab, mode) == CODE_FOR_nothing) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Aligned load, but unsupported type."); return false; @@ -4472,7 +4472,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat &def_stmt, &def, &gather_dt, &gather_off_vectype)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "gather index use not simple."); return false; @@ -4492,7 +4492,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat size_zero_node) < 0; if (negative && ncopies > 1) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "multiple types with negative step."); return false; @@ -4505,14 +4505,14 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat if (alignment_support_scheme != dr_aligned && alignment_support_scheme != dr_unaligned_supported) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "negative step but alignment required."); return false; } if (!perm_mask_for_reverse (vectype)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "negative step and reversing not supported."); return false; @@ -4527,7 +4527,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat return true; } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "transform load. ncopies = %d", ncopies); @@ -5334,7 +5334,7 @@ vectorizable_condition (gimple stmt, gimple_stmt_i /* FORNOW: not yet supported. */ if (STMT_VINFO_LIVE_P (stmt_info)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "value used after loop."); return false; @@ -5534,7 +5534,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect gimple pattern_stmt; gimple_seq pattern_def_seq; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "==> examining statement: "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, stmt, 0); @@ -5542,7 +5542,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect if (gimple_has_volatile_ops (stmt)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: stmt has volatile operands"); @@ -5575,7 +5575,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect /* Analyze PATTERN_STMT instead of the original stmt. */ stmt = pattern_stmt; stmt_info = vinfo_for_stmt (pattern_stmt); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "==> examining pattern statement: "); @@ -5584,7 +5584,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect } else { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "irrelevant."); return true; @@ -5597,7 +5597,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect || STMT_VINFO_LIVE_P (vinfo_for_stmt (pattern_stmt)))) { /* Analyze PATTERN_STMT too. */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "==> examining pattern statement: "); @@ -5621,7 +5621,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect || STMT_VINFO_LIVE_P (vinfo_for_stmt (pattern_def_stmt))) { /* Analyze def stmt of STMT if it's a pattern stmt. */ - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "==> examining pattern def statement: "); @@ -5660,7 +5660,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect gcc_assert (PURE_SLP_STMT (stmt_info)); scalar_type = TREE_TYPE (gimple_get_lhs (stmt)); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "get vectype for scalar type: "); @@ -5670,7 +5670,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect vectype = get_vectype_for_scalar_type (scalar_type); if (!vectype) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not SLPed: unsupported data-type "); @@ -5680,7 +5680,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect return false; } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "vectype: "); dump_generic_expr (MSG_NOTE, TDF_SLIM, vectype); @@ -5724,7 +5724,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect if (!ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: relevant stmt not "); @@ -5746,7 +5746,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect if (!ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not vectorized: live stmt not "); @@ -5846,7 +5846,7 @@ vect_transform_stmt (gimple stmt, gimple_stmt_iter default: if (!STMT_VINFO_LIVE_P (stmt_info)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "stmt not supported."); gcc_unreachable (); @@ -5871,7 +5871,7 @@ vect_transform_stmt (gimple stmt, gimple_stmt_iter tree scalar_dest; gimple exit_phi; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "Record the vdef for outer-loop vectorization."); @@ -6108,7 +6108,7 @@ get_vectype_for_scalar_type_and_size (tree scalar_ return NULL_TREE; vectype = build_vector_type (scalar_type, nunits); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "get vectype with %d units of type ", nunits); @@ -6118,7 +6118,7 @@ get_vectype_for_scalar_type_and_size (tree scalar_ if (!vectype) return NULL_TREE; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "vectype: "); dump_generic_expr (MSG_NOTE, TDF_SLIM, vectype); @@ -6127,7 +6127,7 @@ get_vectype_for_scalar_type_and_size (tree scalar_ if (!VECTOR_MODE_P (TYPE_MODE (vectype)) && !INTEGRAL_MODE_P (TYPE_MODE (vectype))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "mode not supported by target."); return NULL_TREE; @@ -6198,7 +6198,7 @@ vect_is_simple_use (tree operand, gimple stmt, loo *def_stmt = NULL; *def = NULL_TREE; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "vect_is_simple_use: operand "); @@ -6220,14 +6220,14 @@ vect_is_simple_use (tree operand, gimple stmt, loo if (TREE_CODE (operand) == PAREN_EXPR) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "non-associatable copy."); operand = TREE_OPERAND (operand, 0); } if (TREE_CODE (operand) != SSA_NAME) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "not ssa-name."); return false; @@ -6236,13 +6236,13 @@ vect_is_simple_use (tree operand, gimple stmt, loo *def_stmt = SSA_NAME_DEF_STMT (operand); if (*def_stmt == NULL) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "no def_stmt."); return false; } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "def_stmt: "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, *def_stmt, 0); @@ -6274,13 +6274,13 @@ vect_is_simple_use (tree operand, gimple stmt, loo && *dt == vect_double_reduction_def && gimple_code (stmt) != GIMPLE_PHI)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Unsupported pattern."); return false; } - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "type of def: %d.", *dt); switch (gimple_code (*def_stmt)) @@ -6299,7 +6299,7 @@ vect_is_simple_use (tree operand, gimple stmt, loo break; /* FALLTHRU */ default: - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "unsupported defining stmt: "); return false; Index: tree-vect-slp.c =================================================================== --- tree-vect-slp.c (revision 192695) +++ tree-vect-slp.c (working copy) @@ -238,7 +238,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi &def, &dt) || (!def_stmt && dt != vect_constant_def)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: can't find def for "); @@ -263,7 +263,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi pattern = true; if (!first && !oprnd_info->first_pattern) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: some of the stmts" @@ -279,7 +279,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi if (dt == vect_unknown_def_type) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Unsupported pattern."); return false; @@ -296,7 +296,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi break; default: - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "unsupported defining stmt: "); return false; @@ -361,7 +361,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi { if (number_of_oprnds != 2) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: different types "); @@ -388,7 +388,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi && !types_compatible_p (oprnd_info->first_def_type, TREE_TYPE (def_op0)))) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Swapping operands of "); @@ -400,7 +400,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi } else { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: different types "); @@ -435,7 +435,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi default: /* FORNOW: Not supported. */ - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: illegal type of def "); @@ -504,7 +504,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ /* For every stmt in NODE find its def stmt/s. */ FOR_EACH_VEC_ELT (gimple, stmts, i, stmt) { - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Build SLP for "); dump_gimple_stmt (MSG_NOTE, TDF_SLIM, stmt, 0); @@ -513,7 +513,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ /* Fail to vectorize statements marked as unvectorizable. */ if (!STMT_VINFO_VECTORIZABLE (vinfo_for_stmt (stmt))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: unvectorizable statement "); @@ -527,7 +527,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ lhs = gimple_get_lhs (stmt); if (lhs == NULL_TREE) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: not GIMPLE_ASSIGN nor " @@ -544,7 +544,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ && (cond = gimple_assign_rhs1 (stmt)) && !COMPARISON_CLASS_P (cond)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: condition is not " @@ -560,7 +560,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ vectype = get_vectype_for_scalar_type (scalar_type); if (!vectype) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: unsupported data-type "); @@ -591,7 +591,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ || !gimple_call_nothrow_p (stmt) || gimple_call_chain (stmt)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: unsupported call type "); @@ -631,7 +631,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ if (!optab) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: no optab."); vect_free_oprnd_info (&oprnds_info); @@ -640,7 +640,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ icode = (int) optab_handler (optab, vec_mode); if (icode == CODE_FOR_nothing) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: " "op not supported by target."); @@ -674,7 +674,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ || first_stmt_code == COMPONENT_REF || first_stmt_code == MEM_REF))) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: different operation " @@ -689,7 +689,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ if (need_same_oprnds && !operand_equal_p (first_op1, gimple_assign_rhs2 (stmt), 0)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: different shift " @@ -710,7 +710,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ || gimple_call_fntype (first_stmt) != gimple_call_fntype (stmt)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: different calls in "); @@ -749,7 +749,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ || (GROUP_FIRST_ELEMENT (vinfo_for_stmt (stmt)) != stmt && GROUP_GAP (vinfo_for_stmt (stmt)) != 1)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: grouped " @@ -767,7 +767,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ if (loop_vinfo && GROUP_SIZE (vinfo_for_stmt (stmt)) > ncopies * group_size) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: the number " @@ -792,7 +792,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ && rhs_code != REALPART_EXPR && rhs_code != IMAGPART_EXPR) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, @@ -817,7 +817,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ if (vect_supportable_dr_alignment (first_dr, false) == dr_unaligned_unsupported) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, @@ -857,7 +857,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ if (TREE_CODE_CLASS (rhs_code) == tcc_reference) { /* Not grouped load. */ - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: not grouped load "); @@ -875,7 +875,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ && rhs_code != COND_EXPR && rhs_code != CALL_EXPR) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: operation"); @@ -895,7 +895,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_ first_cond_code = TREE_CODE (cond_expr); else if (first_cond_code != TREE_CODE (cond_expr)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: different" @@ -1080,7 +1080,7 @@ vect_supported_slp_permutation_p (slp_instance ins /* Check that the loads are all in the same interleaving chain. */ if (GROUP_FIRST_ELEMENT (vinfo_for_stmt (scalar_stmt)) != first_load) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: unsupported data " @@ -1169,7 +1169,7 @@ vect_supported_load_permutation_p (slp_instance sl if (!slp_instn) return false; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_NOTE, vect_location, "Load permutation "); FOR_EACH_VEC_ELT (int, load_permutation, i, next) @@ -1376,7 +1376,7 @@ vect_supported_load_permutation_p (slp_instance sl if (vect_supportable_dr_alignment (dr, false) == dr_unaligned_unsupported) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, @@ -1536,7 +1536,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf if (!vectype) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: unsupported data-type "); @@ -1556,7 +1556,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf unrolling_factor = least_common_multiple (nunits, group_size) / group_size; if (unrolling_factor != 1 && !loop_vinfo) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: unrolling required in basic" " block SLP"); @@ -1618,7 +1618,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf if (unrolling_factor != 1 && !loop_vinfo) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: unrolling required in basic" " block SLP"); @@ -1645,7 +1645,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf if (!vect_supported_load_permutation_p (new_instance, group_size, load_permutation)) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) { dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Build SLP failed: unsupported load " @@ -1685,7 +1685,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf VEC_safe_push (slp_instance, heap, BB_VINFO_SLP_INSTANCES (bb_vinfo), new_instance); - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) vect_print_slp_tree (MSG_NOTE, node); return true; @@ -1717,7 +1717,7 @@ vect_analyze_slp (loop_vec_info loop_vinfo, bb_vec gimple first_element; bool ok = false; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_analyze_slp ==="); if (loop_vinfo) @@ -1736,7 +1736,7 @@ vect_analyze_slp (loop_vec_info loop_vinfo, bb_vec if (bb_vinfo && !ok) { - if (dump_kind_p (MSG_MISSED_OPTIMIZATION)) + if (dump_enabled_p ()) dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location, "Failed to SLP the basic block."); @@ -1780,7 +1780,7 @@ vect_make_slp_decision (loop_vec_info loop_vinfo) slp_instance instance; int decided_to_slp = 0; - if (dump_kind_p (MSG_NOTE)) + if (dump_enabled_p ()) dump_printf_loc (MSG_NOTE, vect_location, "=== vect_make_slp_decision ===");