[PATCH] vectorization passes clean up for dump info

Richard Biener richard.guenther@gmail.com
Wed Oct 24 10:17:00 GMT 2012


On Wed, Oct 24, 2012 at 2:53 AM, Sharad Singhai <singhai@google.com> wrote:
> Okay, I have updated the patch (attached) to allow inlining of
> dump_enabled_p. I needed to make alt_dump_file extern to for this
> (dump_file was already extern). Also, I moved dump_file definition
> from toplev.c to dumpfile.c, where it seems to logically belong.
>
> Bootstrapped and tested on x86_64. Okay for trunk?

Ok!

Thanks for keeping up with our changing minds ;)

Richard.

> Thanks,
> Sharad
>
> 2012-10-24  Sharad Singhai  <singhai@google.com>
>
>         * dumpfile.c (dump_enabled_p): Make it inline and move the definition
>         to dumpfile.h.
>         (dump_kind_p): Deleted. Functionality replaced by dump_enabled_p.
>         Make alt_dump_file extern.
>         * dumpfile.h (dump_enabled_p): Move inline definition here.
>         (dump_kind_p): Delete declaration.
>         Add extern declaration of alt_dump_file.
>         * toplev.c: Move dump_file and dump_file_name to dumpfile.c.
>         * tree-vect-loop-manip.c: Replace all uses of dump_kind_p with
>         dump_enabled_p.
>         * tree-vectorizer.c: Likewise.
>         * tree-vect-loop.c: Likewise.
>         * tree-vect-data-refs.c: Likewise.
>         * tree-vect-patterns.c: Likewise.
>         * tree-vect-stmts.c: Likewise.
>         * tree-vect-slp.c: Likewise.
>
> Index: dumpfile.c
> ===================================================================
> --- dumpfile.c (revision 192695)
> +++ dumpfile.c (working copy)
> @@ -32,12 +32,18 @@ along with GCC; see the file COPYING3.  If not see
>
>  static int pflags;                   /* current dump_flags */
>  static int alt_flags;                /* current opt_info flags */
> -static FILE *alt_dump_file = NULL;
>
>  static void dump_loc (int, FILE *, source_location);
>  static int dump_phase_enabled_p (int);
>  static FILE *dump_open_alternate_stream (struct dump_file_info *);
>
> +/* These are currently used for communicating between passes.
> +   However, instead of accessing them directly, the passes can use
> +   dump_printf () for dumps.  */
> +FILE *dump_file = NULL;
> +FILE *alt_dump_file = NULL;
> +const char *dump_file_name;
> +
>  /* Table of tree dump switches. This must be consistent with the
>     TREE_DUMP_INDEX enumeration in dumpfile.h.  */
>  static struct dump_file_info dump_files[TDI_end] =
> @@ -514,14 +520,6 @@ dump_phase_enabled_p (int phase)
>      }
>  }
>
> -/* Return true if any of the dumps are enabled, false otherwise. */
> -
> -inline bool
> -dump_enabled_p (void)
> -{
> -  return (dump_file || alt_dump_file);
> -}
> -
>  /* Returns nonzero if tree dump PHASE has been initialized.  */
>
>  int
> @@ -836,16 +834,6 @@ opt_info_switch_p (const char *arg)
>    return opt_info_enable_all ((TDF_TREE | TDF_RTL | TDF_IPA), flags, filename);
>  }
>
> -/* Return true if any dumps are enabled for the given MSG_TYPE, false
> -   otherwise.  */
> -
> -bool
> -dump_kind_p (int msg_type)
> -{
> -  return (dump_file && (msg_type & pflags))
> -    || (alt_dump_file && (msg_type & alt_flags));
> -}
> -
>  /* Print basic block on the dump streams.  */
>
>  void
> Index: dumpfile.h
> ===================================================================
> --- dumpfile.h (revision 192695)
> +++ dumpfile.h (working copy)
> @@ -108,7 +108,6 @@ struct dump_file_info
>    int num;                      /* dump file number */
>  };
>
> -
>  /* In dumpfile.c */
>  extern char *get_dump_file_name (int);
>  extern int dump_initialized_p (int);
> @@ -120,8 +119,6 @@ extern void dump_node (const_tree, int, FILE *);
>  extern int dump_switch_p (const char *);
>  extern int opt_info_switch_p (const char *);
>  extern const char *dump_flag_name (int);
> -extern bool dump_kind_p (int);
> -extern inline bool dump_enabled_p (void);
>  extern void dump_printf (int, const char *, ...) ATTRIBUTE_PRINTF_2;
>  extern void dump_printf_loc (int, source_location,
>                               const char *, ...) ATTRIBUTE_PRINTF_3;
> @@ -142,10 +139,19 @@ extern void dump_bb (FILE *, basic_block, int, int
>
>  /* Global variables used to communicate with passes.  */
>  extern FILE *dump_file;
> +extern FILE *alt_dump_file;
>  extern int dump_flags;
>  extern const char *dump_file_name;
>
>  /* Return the dump_file_info for the given phase.  */
>  extern struct dump_file_info *get_dump_file_info (int);
>
> +/* Return true if any of the dumps are enabled, false otherwise. */
> +
> +static inline bool
> +dump_enabled_p (void)
> +{
> +  return (dump_file || alt_dump_file);
> +}
> +
>  #endif /* GCC_DUMPFILE_H */
> Index: toplev.c
> ===================================================================
> --- toplev.c (revision 192695)
> +++ toplev.c (working copy)
> @@ -169,8 +169,6 @@ const char *user_label_prefix;
>  FILE *asm_out_file;
>  FILE *aux_info_file;
>  FILE *stack_usage_file = NULL;
> -FILE *dump_file = NULL;
> -const char *dump_file_name;
>
>  /* The current working directory of a translation.  It's generally the
>     directory from which compilation was initiated, but a preprocessed
> Index: tree-vect-loop-manip.c
> ===================================================================
> --- tree-vect-loop-manip.c (revision 192695)
> +++ tree-vect-loop-manip.c (working copy)
> @@ -792,7 +792,7 @@ slpeel_make_loop_iterate_ntimes (struct loop *loop
>    free_stmt_vec_info (orig_cond);
>
>    loop_loc = find_loop_location (loop);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        if (LOCATION_LOCUS (loop_loc) != UNKNOWN_LOC)
>   dump_printf (MSG_NOTE, "\nloop at %s:%d: ", LOC_FILE (loop_loc),
> @@ -1683,7 +1683,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo)
>
>    /* Analyze phi functions of the loop header.  */
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "vect_can_advance_ivs_p:");
>    for (gsi = gsi_start_phis (bb); !gsi_end_p (gsi); gsi_next (&gsi))
>      {
> @@ -1691,7 +1691,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo)
>        tree evolution_part;
>
>        phi = gsi_stmt (gsi);
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>            dump_printf_loc (MSG_NOTE, vect_location, "Analyze phi: ");
>            dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0);
> @@ -1702,7 +1702,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo)
>
>        if (virtual_operand_p (PHI_RESULT (phi)))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "virtual phi. skip.");
>    continue;
> @@ -1712,7 +1712,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo)
>
>        if (STMT_VINFO_DEF_TYPE (vinfo_for_stmt (phi)) == vect_reduction_def)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "reduc phi. skip.");
>            continue;
> @@ -1725,13 +1725,13 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo)
>
>        if (!access_fn)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "No Access function.");
>    return false;
>   }
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          {
>    dump_printf_loc (MSG_NOTE, vect_location,
>                             "Access function of PHI: ");
> @@ -1742,7 +1742,7 @@ vect_can_advance_ivs_p (loop_vec_info loop_vinfo)
>
>        if (evolution_part == NULL_TREE)
>          {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf (MSG_MISSED_OPTIMIZATION, "No evolution.");
>    return false;
>          }
> @@ -1827,7 +1827,7 @@ vect_update_ivs_after_vectorizer (loop_vec_info lo
>
>        phi = gsi_stmt (gsi);
>        phi1 = gsi_stmt (gsi1);
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "vect_update_ivs_after_vectorizer: phi: ");
> @@ -1837,7 +1837,7 @@ vect_update_ivs_after_vectorizer (loop_vec_info lo
>        /* Skip virtual phi's.  */
>        if (virtual_operand_p (PHI_RESULT (phi)))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "virtual phi. skip.");
>    continue;
> @@ -1847,7 +1847,7 @@ vect_update_ivs_after_vectorizer (loop_vec_info lo
>        stmt_info = vinfo_for_stmt (phi);
>        if (STMT_VINFO_DEF_TYPE (stmt_info) == vect_reduction_def)
>          {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "reduc phi. skip.");
>            continue;
> @@ -1910,7 +1910,7 @@ vect_do_peeling_for_loop_bound (loop_vec_info loop
>    tree cond_expr = NULL_TREE;
>    gimple_seq cond_expr_stmt_list = NULL;
>
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                       "=== vect_do_peeling_for_loop_bound ===");
>
> @@ -2022,7 +2022,7 @@ vect_gen_niters_for_prolog_loop (loop_vec_info loo
>      {
>        int npeel = LOOP_PEELING_FOR_ALIGNMENT (loop_vinfo);
>
> -      if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                           "known peeling = %d.", npeel);
>
> @@ -2076,7 +2076,7 @@ vect_gen_niters_for_prolog_loop (loop_vec_info loo
>    if (TREE_CODE (loop_niters) != INTEGER_CST)
>      iters = fold_build2 (MIN_EXPR, niters_type, iters, loop_niters);
>
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                         "niters for prolog loop: ");
> @@ -2134,7 +2134,7 @@ vect_update_inits_of_drs (loop_vec_info loop_vinfo
>    VEC (data_reference_p, heap) *datarefs = LOOP_VINFO_DATAREFS (loop_vinfo);
>    struct data_reference *dr;
>
> - if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> + if (dump_enabled_p ())
>      dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                       "=== vect_update_inits_of_dr ===");
>
> @@ -2163,7 +2163,7 @@ vect_do_peeling_for_alignment (loop_vec_info loop_
>    int max_iter;
>    int bound = 0;
>
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                       "=== vect_do_peeling_for_alignment ===");
>
> @@ -2475,7 +2475,7 @@ vect_create_cond_for_alias_checks (loop_vec_info l
>        segment_length_a = vect_vfa_segment_size (dr_a, length_factor);
>        segment_length_b = vect_vfa_segment_size (dr_b, length_factor);
>
> -      if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                             "create runtime check for data references ");
> @@ -2506,7 +2506,7 @@ vect_create_cond_for_alias_checks (loop_vec_info l
>   *cond_expr = part_cond_expr;
>      }
>
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>       "created %u versioning for alias checks.\n",
>       VEC_length (ddr_p, may_alias_ddrs));
> Index: tree-vectorizer.c
> ===================================================================
> --- tree-vectorizer.c (revision 192695)
> +++ tree-vectorizer.c (working copy)
> @@ -107,7 +107,7 @@ vectorize_loops (void)
>   loop_vec_info loop_vinfo;
>   vect_location = find_loop_location (loop);
>          if (LOCATION_LOCUS (vect_location) != UNKNOWN_LOC
> -    && dump_kind_p (MSG_ALL))
> +    && dump_enabled_p ())
>    dump_printf (MSG_ALL, "\nAnalyzing loop at %s:%d\n",
>                         LOC_FILE (vect_location), LOC_LINE (vect_location));
>
> @@ -118,7 +118,7 @@ vectorize_loops (void)
>    continue;
>
>          if (LOCATION_LOCUS (vect_location) != UNKNOWN_LOC
> -    && dump_kind_p (MSG_ALL))
> +    && dump_enabled_p ())
>            dump_printf (MSG_ALL, "\n\nVectorizing loop at %s:%d\n",
>                         LOC_FILE (vect_location), LOC_LINE (vect_location));
>   vect_transform_loop (loop_vinfo);
> @@ -128,8 +128,8 @@ vectorize_loops (void)
>    vect_location = UNKNOWN_LOC;
>
>    statistics_counter_event (cfun, "Vectorized loops", num_vectorized_loops);
> -  if (dump_kind_p (MSG_ALL)
> -      || (num_vectorized_loops > 0 && dump_kind_p (MSG_ALL)))
> +  if (dump_enabled_p ()
> +      || (num_vectorized_loops > 0 && dump_enabled_p ()))
>      dump_printf_loc (MSG_ALL, vect_location,
>                       "vectorized %u loops in function.\n",
>                       num_vectorized_loops);
> @@ -170,7 +170,7 @@ execute_vect_slp (void)
>        if (vect_slp_analyze_bb (bb))
>          {
>            vect_slp_transform_bb (bb);
> -          if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>       "basic block vectorized using SLP\n");
>          }
> Index: tree-vect-loop.c
> ===================================================================
> --- tree-vect-loop.c (revision 192695)
> +++ tree-vect-loop.c (working copy)
> @@ -187,7 +187,7 @@ vect_determine_vectorization_factor (loop_vec_info
>    gimple_stmt_iterator pattern_def_si = gsi_none ();
>    bool analyze_pattern_stmt = false;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "=== vect_determine_vectorization_factor ===");
>
> @@ -199,7 +199,7 @@ vect_determine_vectorization_factor (loop_vec_info
>   {
>    phi = gsi_stmt (si);
>    stmt_info = vinfo_for_stmt (phi);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "==> examining phi: ");
>        dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0);
> @@ -212,7 +212,7 @@ vect_determine_vectorization_factor (loop_vec_info
>        gcc_assert (!STMT_VINFO_VECTYPE (stmt_info));
>                scalar_type = TREE_TYPE (PHI_RESULT (phi));
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_NOTE, vect_location,
>                                     "get vectype for scalar type:  ");
> @@ -222,7 +222,7 @@ vect_determine_vectorization_factor (loop_vec_info
>        vectype = get_vectype_for_scalar_type (scalar_type);
>        if (!vectype)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                         "not vectorized: unsupported "
> @@ -234,14 +234,14 @@ vect_determine_vectorization_factor (loop_vec_info
>   }
>        STMT_VINFO_VECTYPE (stmt_info) = vectype;
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_NOTE, vect_location, "vectype: ");
>    dump_generic_expr (MSG_NOTE, TDF_SLIM, vectype);
>   }
>
>        nunits = TYPE_VECTOR_SUBPARTS (vectype);
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location, "nunits = %d", nunits);
>
>        if (!vectorization_factor
> @@ -261,7 +261,7 @@ vect_determine_vectorization_factor (loop_vec_info
>
>            stmt_info = vinfo_for_stmt (stmt);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                                 "==> examining statement: ");
> @@ -281,7 +281,7 @@ vect_determine_vectorization_factor (loop_vec_info
>                  {
>                    stmt = pattern_stmt;
>                    stmt_info = vinfo_for_stmt (pattern_stmt);
> -                  if (dump_kind_p (MSG_NOTE))
> +                  if (dump_enabled_p ())
>                      {
>                        dump_printf_loc (MSG_NOTE, vect_location,
>                                         "==> examining pattern statement: ");
> @@ -290,7 +290,7 @@ vect_determine_vectorization_factor (loop_vec_info
>                  }
>                else
>          {
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_NOTE, vect_location, "skip.");
>                    gsi_next (&si);
>            continue;
> @@ -330,7 +330,7 @@ vect_determine_vectorization_factor (loop_vec_info
>
>    if (!gsi_end_p (pattern_def_si))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_NOTE, vect_location,
>                                             "==> examining pattern def stmt: ");
> @@ -353,7 +353,7 @@ vect_determine_vectorization_factor (loop_vec_info
>
>    if (gimple_get_lhs (stmt) == NULL_TREE)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                     "not vectorized: irregular stmt.");
> @@ -365,7 +365,7 @@ vect_determine_vectorization_factor (loop_vec_info
>
>    if (VECTOR_MODE_P (TYPE_MODE (gimple_expr_type (stmt))))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                     "not vectorized: vector stmt in loop:");
> @@ -389,7 +389,7 @@ vect_determine_vectorization_factor (loop_vec_info
>      {
>        gcc_assert (!STMT_VINFO_DATA_REF (stmt_info));
>        scalar_type = TREE_TYPE (gimple_get_lhs (stmt));
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_NOTE, vect_location,
>                                     "get vectype for scalar type:  ");
> @@ -398,7 +398,7 @@ vect_determine_vectorization_factor (loop_vec_info
>        vectype = get_vectype_for_scalar_type (scalar_type);
>        if (!vectype)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                         "not vectorized: unsupported "
> @@ -417,7 +417,7 @@ vect_determine_vectorization_factor (loop_vec_info
>       support one vector size per loop).  */
>    scalar_type = vect_get_smallest_scalar_type (stmt, &dummy,
>         &dummy);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                                 "get vectype for scalar type:  ");
> @@ -426,7 +426,7 @@ vect_determine_vectorization_factor (loop_vec_info
>    vf_vectype = get_vectype_for_scalar_type (scalar_type);
>    if (!vf_vectype)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                     "not vectorized: unsupported data-type ");
> @@ -439,7 +439,7 @@ vect_determine_vectorization_factor (loop_vec_info
>    if ((GET_MODE_SIZE (TYPE_MODE (vectype))
>         != GET_MODE_SIZE (TYPE_MODE (vf_vectype))))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                     "not vectorized: different sized vector "
> @@ -453,14 +453,14 @@ vect_determine_vectorization_factor (loop_vec_info
>        return false;
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "vectype: ");
>        dump_generic_expr (MSG_NOTE, TDF_SLIM, vf_vectype);
>      }
>
>    nunits = TYPE_VECTOR_SUBPARTS (vf_vectype);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "nunits = %d", nunits);
>    if (!vectorization_factor
>        || (nunits > vectorization_factor))
> @@ -475,12 +475,12 @@ vect_determine_vectorization_factor (loop_vec_info
>      }
>
>    /* TODO: Analyze cost. Decide if worth while to vectorize.  */
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "vectorization factor = %d",
>                       vectorization_factor);
>    if (vectorization_factor <= 1)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "not vectorized: unsupported data-type");
>        return false;
> @@ -517,7 +517,7 @@ vect_is_simple_iv_evolution (unsigned loop_nb, tre
>    step_expr = evolution_part;
>    init_expr = unshare_expr (initial_condition_in_loop_num (access_fn,
> loop_nb));
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "step: ");
>        dump_generic_expr (MSG_NOTE, TDF_SLIM, step_expr);
> @@ -530,7 +530,7 @@ vect_is_simple_iv_evolution (unsigned loop_nb, tre
>
>    if (TREE_CODE (step_expr) != INTEGER_CST)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "step unknown.");
>        return false;
> @@ -555,7 +555,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v
>    gimple_stmt_iterator gsi;
>    bool double_reduc;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "=== vect_analyze_scalar_cycles ===");
>
> @@ -569,7 +569,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v
>        tree def = PHI_RESULT (phi);
>        stmt_vec_info stmt_vinfo = vinfo_for_stmt (phi);
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_NOTE, vect_location, "Analyze phi: ");
>    dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0);
> @@ -587,7 +587,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v
>        if (access_fn)
>   {
>    STRIP_NOPS (access_fn);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                                 "Access function of PHI: ");
> @@ -606,7 +606,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v
>
>        gcc_assert (STMT_VINFO_LOOP_PHI_EVOLUTION_PART (stmt_vinfo) !=
> NULL_TREE);
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location, "Detected induction.");
>        STMT_VINFO_DEF_TYPE (stmt_vinfo) = vect_induction_def;
>      }
> @@ -621,7 +621,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v
>        gimple reduc_stmt;
>        bool nested_cycle;
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_NOTE, vect_location, "Analyze phi: ");
>            dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0);
> @@ -637,7 +637,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v
>          {
>            if (double_reduc)
>              {
> -              if (dump_kind_p (MSG_NOTE))
> +              if (dump_enabled_p ())
>                  dump_printf_loc (MSG_NOTE, vect_location,
>   "Detected double reduction.");
>
> @@ -649,7 +649,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v
>              {
>                if (nested_cycle)
>                  {
> -                  if (dump_kind_p (MSG_NOTE))
> +                  if (dump_enabled_p ())
>                      dump_printf_loc (MSG_NOTE, vect_location,
>       "Detected vectorizable nested cycle.");
>
> @@ -659,7 +659,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v
>                  }
>                else
>                  {
> -                  if (dump_kind_p (MSG_NOTE))
> +                  if (dump_enabled_p ())
>                      dump_printf_loc (MSG_NOTE, vect_location,
>       "Detected reduction.");
>
> @@ -675,7 +675,7 @@ vect_analyze_scalar_cycles_1 (loop_vec_info loop_v
>              }
>          }
>        else
> -        if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +        if (dump_enabled_p ())
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "Unknown def-use cycle pattern.");
>      }
> @@ -737,7 +737,7 @@ vect_get_loop_niters (struct loop *loop, tree *num
>  {
>    tree niters;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>       "=== get_loop_niters ===");
>    niters = number_of_exit_cond_executions (loop);
> @@ -747,7 +747,7 @@ vect_get_loop_niters (struct loop *loop, tree *num
>      {
>        *number_of_iterations = niters;
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_NOTE, vect_location, "==> get_loop_niters:");
>            dump_generic_expr (MSG_NOTE, TDF_SLIM, *number_of_iterations);
> @@ -995,7 +995,7 @@ vect_analyze_loop_1 (struct loop *loop)
>  {
>    loop_vec_info loop_vinfo;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>       "===== analyze_loop_nest_1 =====");
>
> @@ -1004,7 +1004,7 @@ vect_analyze_loop_1 (struct loop *loop)
>    loop_vinfo = vect_analyze_loop_form (loop);
>    if (!loop_vinfo)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "bad inner-loop form.");
>        return NULL;
> @@ -1030,7 +1030,7 @@ vect_analyze_loop_form (struct loop *loop)
>    tree number_of_iterations = NULL;
>    loop_vec_info inner_loop_vinfo = NULL;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>       "=== vect_analyze_loop_form ===");
>
> @@ -1054,7 +1054,7 @@ vect_analyze_loop_form (struct loop *loop)
>
>        if (loop->num_nodes != 2)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "not vectorized: control flow in loop.");
>            return NULL;
> @@ -1062,7 +1062,7 @@ vect_analyze_loop_form (struct loop *loop)
>
>        if (empty_block_p (loop->header))
>      {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "not vectorized: empty loop.");
>        return NULL;
> @@ -1092,7 +1092,7 @@ vect_analyze_loop_form (struct loop *loop)
>
>        if ((loop->inner)->inner || (loop->inner)->next)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "not vectorized: multiple nested loops.");
>    return NULL;
> @@ -1102,7 +1102,7 @@ vect_analyze_loop_form (struct loop *loop)
>        inner_loop_vinfo = vect_analyze_loop_1 (loop->inner);
>        if (!inner_loop_vinfo)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>              dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "not vectorized: Bad inner loop.");
>    return NULL;
> @@ -1111,7 +1111,7 @@ vect_analyze_loop_form (struct loop *loop)
>        if (!expr_invariant_in_loop_p (loop,
>   LOOP_VINFO_NITERS (inner_loop_vinfo)))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "not vectorized: inner-loop count not invariant.");
>    destroy_loop_vec_info (inner_loop_vinfo, true);
> @@ -1120,7 +1120,7 @@ vect_analyze_loop_form (struct loop *loop)
>
>        if (loop->num_nodes != 5)
>          {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "not vectorized: control flow in loop.");
>    destroy_loop_vec_info (inner_loop_vinfo, true);
> @@ -1136,14 +1136,14 @@ vect_analyze_loop_form (struct loop *loop)
>    || !single_exit (innerloop)
>    || single_exit (innerloop)->dest !=  EDGE_PRED (loop->latch, 0)->src)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "not vectorized: unsupported outerloop form.");
>    destroy_loop_vec_info (inner_loop_vinfo, true);
>    return NULL;
>   }
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>   "Considering outer-loop vectorization.");
>      }
> @@ -1151,7 +1151,7 @@ vect_analyze_loop_form (struct loop *loop)
>    if (!single_exit (loop)
>        || EDGE_COUNT (loop->header->preds) != 2)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>            if (!single_exit (loop))
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
> @@ -1172,7 +1172,7 @@ vect_analyze_loop_form (struct loop *loop)
>    if (!empty_block_p (loop->latch)
>          || !gimple_seq_empty_p (phi_nodes (loop->latch)))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: unexpected loop form.");
>        if (inner_loop_vinfo)
> @@ -1187,12 +1187,12 @@ vect_analyze_loop_form (struct loop *loop)
>        if (!(e->flags & EDGE_ABNORMAL))
>   {
>    split_loop_exit_edge (e);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf (MSG_NOTE, "split exit edge.");
>   }
>        else
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "not vectorized: abnormal loop exit edge.");
>    if (inner_loop_vinfo)
> @@ -1204,7 +1204,7 @@ vect_analyze_loop_form (struct loop *loop)
>    loop_cond = vect_get_loop_niters (loop, &number_of_iterations);
>    if (!loop_cond)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: complicated exit condition.");
>        if (inner_loop_vinfo)
> @@ -1214,7 +1214,7 @@ vect_analyze_loop_form (struct loop *loop)
>
>    if (!number_of_iterations)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: number of iterations cannot be "
>   "computed.");
> @@ -1225,7 +1225,7 @@ vect_analyze_loop_form (struct loop *loop)
>
>    if (chrec_contains_undetermined (number_of_iterations))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "Infinite number of iterations.");
>        if (inner_loop_vinfo)
> @@ -1235,7 +1235,7 @@ vect_analyze_loop_form (struct loop *loop)
>
>    if (!NITERS_KNOWN_P (number_of_iterations))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_NOTE, vect_location,
>     "Symbolic number of iterations is ");
> @@ -1244,7 +1244,7 @@ vect_analyze_loop_form (struct loop *loop)
>      }
>    else if (TREE_INT_CST_LOW (number_of_iterations) == 0)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: number of iterations = 0.");
>        if (inner_loop_vinfo)
> @@ -1292,7 +1292,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>    HOST_WIDE_INT estimated_niter;
>    int min_profitable_estimate;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>       "=== vect_analyze_loop_operations ===");
>
> @@ -1328,7 +1328,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>   LOOP_VINFO_SLP_UNROLLING_FACTOR (loop_vinfo));
>
>        LOOP_VINFO_VECT_FACTOR (loop_vinfo) = vectorization_factor;
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>   "Updating vectorization factor to %d ",
>   vectorization_factor);
> @@ -1344,7 +1344,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>            ok = true;
>
>            stmt_info = vinfo_for_stmt (phi);
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_NOTE, vect_location, "examining phi: ");
>                dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0);
> @@ -1363,7 +1363,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>                    && STMT_VINFO_DEF_TYPE (stmt_info)
>                       != vect_double_reduction_def)
>                  {
> -                  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "Unsupported loop-closed phi in "
>       "outer-loop.");
> @@ -1405,7 +1405,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>            if (STMT_VINFO_LIVE_P (stmt_info))
>              {
>                /* FORNOW: not yet supported.  */
> -              if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +              if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: value used after loop.");
>                return false;
> @@ -1415,7 +1415,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>                && STMT_VINFO_DEF_TYPE (stmt_info) != vect_induction_def)
>              {
>                /* A scalar-dependence cycle that we don't support.  */
> -              if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +              if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: scalar dependence cycle.");
>                return false;
> @@ -1430,7 +1430,7 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>
>            if (!ok)
>              {
> -              if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +              if (dump_enabled_p ())
>                  {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "not vectorized: relevant phi not "
> @@ -1456,18 +1456,17 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>       touching this loop.  */
>    if (!need_to_vectorize)
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>   "All the computation can be taken out of the loop.");
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: redundant loop. no profit to "
>   "vectorize.");
>        return false;
>      }
>
> -  if (LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo)
> -      && dump_kind_p (MSG_NOTE))
> +  if (LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo) && dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>       "vectorization_factor = %d, niters = "
>       HOST_WIDE_INT_PRINT_DEC, vectorization_factor,
> @@ -1478,10 +1477,10 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>        || ((max_niter = max_stmt_executions_int (loop)) != -1
>    && (unsigned HOST_WIDE_INT) max_niter < vectorization_factor))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: iteration count too small.");
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: iteration count smaller than "
>   "vectorization factor.");
> @@ -1500,10 +1499,10 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>
>    if (min_profitable_iters < 0)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: vectorization not profitable.");
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: vector version will never be "
>   "profitable.");
> @@ -1526,10 +1525,10 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>    if (LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo)
>        && LOOP_VINFO_INT_NITERS (loop_vinfo) <= th)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: vectorization not profitable.");
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>   "not vectorized: iteration count smaller than user "
>   "specified loop bound parameter or minimum profitable "
> @@ -1541,11 +1540,11 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>        && ((unsigned HOST_WIDE_INT) estimated_niter
>            <= MAX (th, (unsigned)min_profitable_estimate)))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: estimated iteration count too "
>                           "small.");
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>   "not vectorized: estimated iteration count smaller "
>                           "than specified loop bound parameter or minimum "
> @@ -1558,18 +1557,18 @@ vect_analyze_loop_operations (loop_vec_info loop_v
>        || LOOP_VINFO_INT_NITERS (loop_vinfo) % vectorization_factor != 0
>        || LOOP_PEELING_FOR_ALIGNMENT (loop_vinfo))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location, "epilog loop required.");
>        if (!vect_can_advance_ivs_p (loop_vinfo))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "not vectorized: can't create epilog loop 1.");
>            return false;
>          }
>        if (!slpeel_can_duplicate_loop_p (loop, single_exit (loop)))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "not vectorized: can't create epilog loop 2.");
>            return false;
> @@ -1602,7 +1601,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo)
>    ok = vect_analyze_data_refs (loop_vinfo, NULL, &min_vf);
>    if (!ok)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "bad data references.");
>        return false;
> @@ -1620,7 +1619,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo)
>    ok = vect_mark_stmts_to_be_vectorized (loop_vinfo);
>    if (!ok)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "unexpected pattern.");
>        return false;
> @@ -1635,7 +1634,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo)
>    if (!ok
>        || max_vf < min_vf)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "bad data dependence.");
>        return false;
> @@ -1644,14 +1643,14 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo)
>    ok = vect_determine_vectorization_factor (loop_vinfo);
>    if (!ok)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "can't determine vectorization factor.");
>        return false;
>      }
>    if (max_vf < LOOP_VINFO_VECT_FACTOR (loop_vinfo))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "bad data dependence.");
>        return false;
> @@ -1663,7 +1662,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo)
>    ok = vect_analyze_data_refs_alignment (loop_vinfo, NULL);
>    if (!ok)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "bad data alignment.");
>        return false;
> @@ -1675,7 +1674,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo)
>    ok = vect_analyze_data_ref_accesses (loop_vinfo, NULL);
>    if (!ok)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "bad data access.");
>        return false;
> @@ -1687,7 +1686,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo)
>    ok = vect_prune_runtime_alias_test_list (loop_vinfo);
>    if (!ok)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "too long list of versioning for alias "
>   "run-time tests.");
> @@ -1700,7 +1699,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo)
>    ok = vect_enhance_data_refs_alignment (loop_vinfo);
>    if (!ok)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "bad data alignment.");
>        return false;
> @@ -1725,7 +1724,7 @@ vect_analyze_loop_2 (loop_vec_info loop_vinfo)
>    ok = vect_analyze_loop_operations (loop_vinfo, slp);
>    if (!ok)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "bad operation or unsupported loop bound.");
>        return false;
> @@ -1749,7 +1748,7 @@ vect_analyze_loop (struct loop *loop)
>    current_vector_size = 0;
>    vector_sizes = targetm.vectorize.autovectorize_vector_sizes ();
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>       "===== analyze_loop_nest =====");
>
> @@ -1757,7 +1756,7 @@ vect_analyze_loop (struct loop *loop)
>        && loop_vec_info_for_loop (loop_outer (loop))
>        && LOOP_VINFO_VECTORIZABLE_P (loop_vec_info_for_loop
> (loop_outer (loop))))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>   "outer-loop already vectorized.");
>        return NULL;
> @@ -1769,7 +1768,7 @@ vect_analyze_loop (struct loop *loop)
>        loop_vinfo = vect_analyze_loop_form (loop);
>        if (!loop_vinfo)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "bad loop form.");
>    return NULL;
> @@ -1791,7 +1790,7 @@ vect_analyze_loop (struct loop *loop)
>
>        /* Try the next biggest vector size.  */
>        current_vector_size = 1 << floor_log2 (vector_sizes);
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>   "***** Re-trying analysis with "
>   "vector size %d\n", current_vector_size);
> @@ -2023,7 +2022,7 @@ vect_is_slp_reduction (loop_vec_info loop_info, gi
>                                    == vect_internal_def
>                        && !is_loop_header_bb_p (gimple_bb (def_stmt)))))
>        {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_NOTE, vect_location, "swapping oprnds: ");
>    dump_gimple_stmt (MSG_NOTE, TDF_SLIM, next_stmt, 0);
> @@ -2125,7 +2124,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>
>        if (!flow_bb_inside_loop_p (loop, gimple_bb (use_stmt)))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "intermediate value used outside loop.");
>
> @@ -2137,7 +2136,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>          nloop_uses++;
>        if (nloop_uses > 1)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "reduction used in loop.");
>            return NULL;
> @@ -2146,7 +2145,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>
>    if (TREE_CODE (loop_arg) != SSA_NAME)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "reduction: not ssa_name: ");
> @@ -2158,7 +2157,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>    def_stmt = SSA_NAME_DEF_STMT (loop_arg);
>    if (!def_stmt)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "reduction: no def_stmt.");
>        return NULL;
> @@ -2166,7 +2165,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>
>    if (!is_gimple_assign (def_stmt) && gimple_code (def_stmt) != GIMPLE_PHI)
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_gimple_stmt (MSG_NOTE, TDF_SLIM, def_stmt, 0);
>        return NULL;
>      }
> @@ -2194,7 +2193,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>   nloop_uses++;
>        if (nloop_uses > 1)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "reduction used in loop.");
>    return NULL;
> @@ -2210,7 +2209,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>        if (gimple_phi_num_args (def_stmt) != 1
>            || TREE_CODE (op1) != SSA_NAME)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "unsupported phi node definition.");
>
> @@ -2223,7 +2222,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>            && flow_bb_inside_loop_p (loop->inner, gimple_bb (def1))
>            && is_gimple_assign (def1))
>          {
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              report_vect_op (MSG_NOTE, def_stmt,
>      "detected double reduction: ");
>
> @@ -2250,7 +2249,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>    if (check_reduction
>        && (!commutative_tree_code (code) || !associative_tree_code (code)))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt,
>   "reduction: not commutative/associative: ");
>        return NULL;
> @@ -2260,7 +2259,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>      {
>        if (code != COND_EXPR)
>          {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt,
>      "reduction: not binary operation: ");
>
> @@ -2279,7 +2278,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>
>        if (TREE_CODE (op1) != SSA_NAME && TREE_CODE (op2) != SSA_NAME)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt,
>      "reduction: uses not ssa_names: ");
>
> @@ -2293,7 +2292,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>
>        if (TREE_CODE (op1) != SSA_NAME && TREE_CODE (op2) != SSA_NAME)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt,
>      "reduction: uses not ssa_names: ");
>
> @@ -2311,7 +2310,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>        || (op4 && TREE_CODE (op4) == SSA_NAME
>            && !types_compatible_p (type, TREE_TYPE (op4))))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_NOTE, vect_location,
>     "reduction: multiple types: operation type: ");
> @@ -2353,7 +2352,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>        && check_reduction)
>      {
>        /* Changing the order of operations changes the semantics.  */
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt,
>   "reduction: unsafe fp math optimization: ");
>        return NULL;
> @@ -2362,7 +2361,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>     && check_reduction)
>      {
>        /* Changing the order of operations changes the semantics.  */
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt,
>   "reduction: unsafe int math optimization: ");
>        return NULL;
> @@ -2370,7 +2369,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>    else if (SAT_FIXED_POINT_TYPE_P (type) && check_reduction)
>      {
>        /* Changing the order of operations changes the semantics.  */
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt,
>   "reduction: unsafe fixed-point math optimization: ");
>        return NULL;
> @@ -2407,7 +2406,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>    if (code != COND_EXPR
>        && ((!def1 || gimple_nop_p (def1)) && (!def2 || gimple_nop_p (def2))))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   report_vect_op (MSG_NOTE, def_stmt, "reduction: no defs for operands: ");
>        return NULL;
>      }
> @@ -2429,7 +2428,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>                            == vect_internal_def
>                && !is_loop_header_bb_p (gimple_bb (def1)))))))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   report_vect_op (MSG_NOTE, def_stmt, "detected reduction: ");
>        return def_stmt;
>      }
> @@ -2452,7 +2451,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>            /* Swap operands (just for simplicity - so that the rest of the code
>       can assume that the reduction variable is always the last (second)
>       argument).  */
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>      report_vect_op (MSG_NOTE, def_stmt,
>                "detected reduction: need to swap operands: ");
>
> @@ -2464,7 +2463,7 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>          }
>        else
>          {
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              report_vect_op (MSG_NOTE, def_stmt, "detected reduction: ");
>          }
>
> @@ -2474,14 +2473,14 @@ vect_is_simple_reduction_1 (loop_vec_info loop_inf
>    /* Try to find SLP reduction chain.  */
>    if (check_reduction && vect_is_slp_reduction (loop_info, phi, def_stmt))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          report_vect_op (MSG_NOTE, def_stmt,
>   "reduction: detected reduction chain: ");
>
>        return def_stmt;
>      }
>
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      report_vect_op (MSG_MISSED_OPTIMIZATION, def_stmt,
>      "reduction: unknown pattern: ");
>
> @@ -2589,7 +2588,7 @@ vect_get_known_peeling_cost (loop_vec_info loop_vi
>    if (!LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo))
>      {
>        *peel_iters_epilogue = vf/2;
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>   "cost model: epilogue peel iters set to vf/2 "
>   "because loop iterations are unknown .");
> @@ -2882,7 +2881,7 @@ vect_estimate_min_profitable_iters (loop_vec_info
>    /* vector version will never be profitable.  */
>    else
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "cost model: the vector iteration cost = %d "
>   "divided by the scalar iteration cost = %d "
> @@ -2893,7 +2892,7 @@ vect_estimate_min_profitable_iters (loop_vec_info
>        return;
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "Cost model analysis: \n");
>        dump_printf (MSG_NOTE, "  Vector inside of loop cost: %d\n",
> @@ -2925,7 +2924,7 @@ vect_estimate_min_profitable_iters (loop_vec_info
>         then skip the vectorized loop.  */
>    min_profitable_iters--;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "  Runtime profitability threshold = %d\n",
> min_profitable_iters);
>
> @@ -2950,7 +2949,7 @@ vect_estimate_min_profitable_iters (loop_vec_info
>      }
>    min_profitable_estimate --;
>    min_profitable_estimate = MAX (min_profitable_estimate,
> min_profitable_iters);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "  Static estimate profitability threshold = %d\n",
>                        min_profitable_iters);
> @@ -3010,7 +3009,7 @@ vect_model_reduction_cost (stmt_vec_info stmt_info
>    vectype = get_vectype_for_scalar_type (TREE_TYPE (reduction_op));
>    if (!vectype)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "unsupported data-type ");
> @@ -3081,7 +3080,7 @@ vect_model_reduction_cost (stmt_vec_info stmt_info
>   }
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf (MSG_NOTE,
>                   "vect_model_reduction_cost: inside_cost = %d, "
>                   "prologue_cost = %d, epilogue_cost = %d .", inside_cost,
> @@ -3110,7 +3109,7 @@ vect_model_induction_cost (stmt_vec_info stmt_info
>    prologue_cost = add_stmt_cost (target_cost_data, 2, scalar_to_vec,
>   stmt_info, 0, vect_prologue);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "vect_model_induction_cost: inside_cost = %d, "
>                       "prologue_cost = %d .", inside_cost, prologue_cost);
> @@ -3239,7 +3238,7 @@ get_initial_def_for_induction (gimple iv_phi)
>    new_bb = gsi_insert_on_edge_immediate (pe, init_stmt);
>    gcc_assert (!new_bb);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>         "created new init_stmt: ");
> @@ -3382,7 +3381,7 @@ get_initial_def_for_induction (gimple iv_phi)
>        && !STMT_VINFO_LIVE_P (stmt_vinfo));
>
>    STMT_VINFO_VEC_STMT (stmt_vinfo) = new_stmt;
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>         "vector of inductions after inner-loop:");
> @@ -3392,7 +3391,7 @@ get_initial_def_for_induction (gimple iv_phi)
>      }
>
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>         "transform induction: created def-use cycle: ");
> @@ -3800,7 +3799,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap)
>
>            add_phi_arg (phi, def, loop_latch_edge (loop), UNKNOWN_LOCATION);
>
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_NOTE, vect_location,
>         "transform reduction: created def-use cycle: ");
> @@ -4001,7 +4000,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap)
>        /*** Case 1:  Create:
>             v_out2 = reduc_expr <v_out1>  */
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>   "Reduce using direct vector reduction.");
>
> @@ -4052,7 +4051,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap)
>                    Create:  va = vop <va, va'>
>                  }  */
>
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_NOTE, vect_location,
>       "Reduce using vector shifts");
>
> @@ -4093,7 +4092,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap)
>                   Create:  s = op <s, s'>  // For non SLP cases
>                 }  */
>
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_NOTE, vect_location,
>       "Reduce using scalar code. ");
>
> @@ -4184,7 +4183,7 @@ vect_create_epilog_for_reduction (VEC (tree, heap)
>      {
>        tree rhs;
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>   "extract scalar result");
>
> @@ -4423,7 +4422,7 @@ vect_finalize_reduction:
>                                 UNKNOWN_LOCATION);
>                    add_phi_arg (vect_phi, PHI_RESULT (inner_phi),
>                                 loop_latch_edge (outer_loop), UNKNOWN_LOCATION);
> -                  if (dump_kind_p (MSG_NOTE))
> +                  if (dump_enabled_p ())
>                      {
>                        dump_printf_loc (MSG_NOTE, vect_location,
>         "created double reduction phi node: ");
> @@ -4773,7 +4772,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i
>      {
>        if (!vectorizable_condition (stmt, gsi, NULL, ops[reduc_index], 0, NULL))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "unsupported condition in reduction");
>
> @@ -4788,7 +4787,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i
>        optab = optab_for_tree_code (code, vectype_in, optab_default);
>        if (!optab)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "no optab.");
>
> @@ -4797,7 +4796,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i
>
>        if (optab_handler (optab, vec_mode) == CODE_FOR_nothing)
>          {
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              dump_printf (MSG_NOTE, "op not supported by target.");
>
>            if (GET_MODE_SIZE (vec_mode) != UNITS_PER_WORD
> @@ -4805,7 +4804,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i
>            < vect_min_worthwhile_factor (code))
>              return false;
>
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>        dump_printf (MSG_NOTE, "proceeding using word mode.");
>          }
>
> @@ -4814,7 +4813,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i
>            && LOOP_VINFO_VECT_FACTOR (loop_vinfo)
>         < vect_min_worthwhile_factor (code))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "not worthwhile without SIMD support.");
>
> @@ -4895,7 +4894,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i
>                                           optab_default);
>        if (!reduc_optab)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "no optab for reduction.");
>
> @@ -4905,7 +4904,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i
>        if (reduc_optab
>            && optab_handler (reduc_optab, vec_mode) == CODE_FOR_nothing)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "reduc op not supported by target.");
>
> @@ -4916,7 +4915,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i
>      {
>        if (!nested_cycle || double_reduc)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "no reduc code for scalar code.");
>
> @@ -4926,7 +4925,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i
>
>    if (double_reduc && ncopies > 1)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "multiple types in double reduction");
>
> @@ -4945,7 +4944,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i
>          ops[1] = fold_convert (TREE_TYPE (ops[0]), ops[1]);
>        else
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "invalid types in dot-prod");
>
> @@ -4963,7 +4962,7 @@ vectorizable_reduction (gimple stmt, gimple_stmt_i
>
>    /** Transform.  **/
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "transform reduction.");
>
>    /* FORNOW: Multiple types are not supported for condition.  */
> @@ -5249,7 +5248,7 @@ vectorizable_induction (gimple phi, gimple_stmt_it
>
>        if (ncopies > 1)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "multiple types in nested loop.");
>    return false;
> @@ -5273,7 +5272,7 @@ vectorizable_induction (gimple phi, gimple_stmt_it
>    if (!(STMT_VINFO_RELEVANT_P (exit_phi_vinfo)
>   && !STMT_VINFO_LIVE_P (exit_phi_vinfo)))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "inner-loop induction only used outside "
>   "of the outer vectorized loop.");
> @@ -5297,7 +5296,7 @@ vectorizable_induction (gimple phi, gimple_stmt_it
>    if (!vec_stmt) /* transformation not required.  */
>      {
>        STMT_VINFO_TYPE (stmt_info) = induc_vec_info_type;
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>                           "=== vectorizable_induction ===");
>        vect_model_induction_cost (stmt_info, ncopies);
> @@ -5306,7 +5305,7 @@ vectorizable_induction (gimple phi, gimple_stmt_it
>
>    /** Transform.  **/
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "transform induction phi.");
>
>    vec_def = get_initial_def_for_induction (phi);
> @@ -5371,7 +5370,7 @@ vectorizable_live_operation (gimple stmt,
>            && !vect_is_simple_use (op, stmt, loop_vinfo, NULL, &def_stmt, &def,
>    &dt))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "use not simple.");
>            return false;
> @@ -5410,7 +5409,7 @@ vect_loop_kill_debug_uses (struct loop *loop, gimp
>      {
>        if (gimple_debug_bind_p (ustmt))
>   {
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                                       "killing debug use");
>
> @@ -5450,7 +5449,7 @@ vect_transform_loop (loop_vec_info loop_vinfo)
>    bool check_profitability = false;
>    int th;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "=== vec_transform_loop ===");
>
>    /* Use the more conservative vectorization threshold.  If the number
> @@ -5464,7 +5463,7 @@ vect_transform_loop (loop_vec_info loop_vinfo)
>    if (th >= LOOP_VINFO_VECT_FACTOR (loop_vinfo) - 1
>        && !LOOP_VINFO_NITERS_KNOWN_P (loop_vinfo))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>   "Profitability threshold is %d loop iterations.", th);
>        check_profitability = true;
> @@ -5525,7 +5524,7 @@ vect_transform_loop (loop_vec_info loop_vinfo)
>        for (si = gsi_start_phis (bb); !gsi_end_p (si); gsi_next (&si))
>          {
>    phi = gsi_stmt (si);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                                 "------>vectorizing phi: ");
> @@ -5544,12 +5543,12 @@ vect_transform_loop (loop_vec_info loop_vinfo)
>
>    if ((TYPE_VECTOR_SUBPARTS (STMT_VINFO_VECTYPE (stmt_info))
>          != (unsigned HOST_WIDE_INT) vectorization_factor)
> -      && dump_kind_p (MSG_NOTE))
> +      && dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "multiple-types.");
>
>    if (STMT_VINFO_DEF_TYPE (stmt_info) == vect_induction_def)
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location, "transform phi.");
>        vect_transform_stmt (phi, NULL, NULL, NULL, NULL);
>      }
> @@ -5565,7 +5564,7 @@ vect_transform_loop (loop_vec_info loop_vinfo)
>            else
>              stmt = gsi_stmt (si);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>         "------>vectorizing statement: ");
> @@ -5637,7 +5636,7 @@ vect_transform_loop (loop_vec_info loop_vinfo)
>
>    if (!gsi_end_p (pattern_def_si))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_NOTE, vect_location,
>     "==> vectorizing pattern def "
> @@ -5664,7 +5663,7 @@ vect_transform_loop (loop_vec_info loop_vinfo)
>                                                 STMT_VINFO_VECTYPE (stmt_info));
>    if (!STMT_SLP_TYPE (stmt_info)
>        && nunits != (unsigned int) vectorization_factor
> -              && dump_kind_p (MSG_NOTE))
> +              && dump_enabled_p ())
>      /* For SLP VF is set according to unrolling factor, and not to
>         vector size, hence for SLP this print is not valid.  */
>              dump_printf_loc (MSG_NOTE, vect_location,
> @@ -5678,7 +5677,7 @@ vect_transform_loop (loop_vec_info loop_vinfo)
>   {
>    slp_scheduled = true;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>       "=== scheduling SLP instances ===");
>
> @@ -5698,7 +5697,7 @@ vect_transform_loop (loop_vec_info loop_vinfo)
>      }
>
>    /* -------- vectorize statement ------------ */
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "transform statement.");
>
>    grouped_store = false;
> @@ -5741,9 +5740,9 @@ vect_transform_loop (loop_vec_info loop_vinfo)
>       until all the loops have been transformed?  */
>    update_ssa (TODO_update_ssa);
>
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location, "LOOP
> VECTORIZED.");
> -  if (loop->inner && dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (loop->inner && dump_enabled_p ())
>      dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>       "OUTER LOOP VECTORIZED.");
>  }
> Index: tree-vect-data-refs.c
> ===================================================================
> --- tree-vect-data-refs.c (revision 192695)
> +++ tree-vect-data-refs.c (working copy)
> @@ -60,7 +60,7 @@ vect_lanes_optab_supported_p (const char *name, co
>
>    if (array_mode == BLKmode)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "no array mode for %s[" HOST_WIDE_INT_PRINT_DEC "]",
>                           GET_MODE_NAME (mode), count);
> @@ -69,14 +69,14 @@ vect_lanes_optab_supported_p (const char *name, co
>
>    if (convert_optab_handler (optab, array_mode, mode) == CODE_FOR_nothing)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "cannot use %s<%s><%s>", name,
>                           GET_MODE_NAME (array_mode), GET_MODE_NAME (mode));
>        return false;
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "can use %s<%s><%s>", name, GET_MODE_NAME (array_mode),
>                       GET_MODE_NAME (mode));
> @@ -439,7 +439,7 @@ vect_check_interleaving (struct data_reference *dr
>        if (diff_mod_size == 0)
>   {
>    vect_update_interleaving_chain (drb, dra);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                                 "Detected interleaving ");
> @@ -462,7 +462,7 @@ vect_check_interleaving (struct data_reference *dr
>        if (diff_mod_size == 0)
>   {
>    vect_update_interleaving_chain (dra, drb);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                                 "Detected interleaving ");
> @@ -524,7 +524,7 @@ vect_mark_for_runtime_alias_test (ddr_p ddr, loop_
>    if ((unsigned) PARAM_VALUE (PARAM_VECT_MAX_VERSION_FOR_ALIAS_CHECKS) == 0)
>      return false;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                         "mark for run-time aliasing test between ");
> @@ -535,7 +535,7 @@ vect_mark_for_runtime_alias_test (ddr_p ddr, loop_
>
>    if (optimize_loop_nest_for_size_p (loop))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "versioning not supported when optimizing for size.");
>        return false;
> @@ -544,7 +544,7 @@ vect_mark_for_runtime_alias_test (ddr_p ddr, loop_
>    /* FORNOW: We don't support versioning with outer-loop vectorization.  */
>    if (loop->inner)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "versioning not yet supported for outer-loops.");
>        return false;
> @@ -555,7 +555,7 @@ vect_mark_for_runtime_alias_test (ddr_p ddr, loop_
>    if (TREE_CODE (DR_STEP (DDR_A (ddr))) != INTEGER_CST
>        || TREE_CODE (DR_STEP (DDR_B (ddr))) != INTEGER_CST)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "versioning not yet supported for non-constant "
>                           "step");
> @@ -611,7 +611,7 @@ vect_analyze_data_ref_dependence (struct data_depe
>
>        if (loop_vinfo)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                 "versioning for alias required: "
> @@ -637,7 +637,7 @@ vect_analyze_data_ref_dependence (struct data_depe
>        if (DR_IS_READ (dra) && DR_IS_READ (drb))
>          return false;
>
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                             "can't determine dependence between ");
> @@ -666,7 +666,7 @@ vect_analyze_data_ref_dependence (struct data_depe
>        if (dra != drb && vect_check_interleaving (dra, drb))
>          return false;
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "determined dependence between ");
> @@ -686,7 +686,7 @@ vect_analyze_data_ref_dependence (struct data_depe
>    /* Loop-based vectorization and known data dependence.  */
>    if (DDR_NUM_DIST_VECTS (ddr) == 0)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                             "versioning for alias required: "
> @@ -704,13 +704,13 @@ vect_analyze_data_ref_dependence (struct data_depe
>      {
>        int dist = dist_v[loop_depth];
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>                           "dependence distance  = %d.", dist);
>
>        if (dist == 0)
>   {
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                                 "dependence distance == 0 between ");
> @@ -737,7 +737,7 @@ vect_analyze_data_ref_dependence (struct data_depe
>    /* If DDR_REVERSED_P the order of the data-refs in DDR was
>       reversed (to make distance vector positive), and the actual
>       distance is negative.  */
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "dependence distance negative.");
>    continue;
> @@ -749,7 +749,7 @@ vect_analyze_data_ref_dependence (struct data_depe
>    /* The dependence distance requires reduction of the maximal
>       vectorization factor.  */
>    *max_vf = abs (dist);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                               "adjusting maximal vectorization factor to %i",
>                               *max_vf);
> @@ -759,13 +759,13 @@ vect_analyze_data_ref_dependence (struct data_depe
>   {
>    /* Dependence distance does not create dependence, as far as
>       vectorization is concerned, in this case.  */
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                               "dependence distance >= VF.");
>    continue;
>   }
>
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                         "not vectorized, possible dependence "
> @@ -795,7 +795,7 @@ vect_analyze_data_ref_dependences (loop_vec_info l
>    VEC (ddr_p, heap) *ddrs = NULL;
>    struct data_dependence_relation *ddr;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "=== vect_analyze_dependences ===");
>    if (loop_vinfo)
> @@ -837,7 +837,7 @@ vect_compute_data_ref_alignment (struct data_refer
>    tree misalign;
>    tree aligned_to, alignment;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "vect_compute_data_ref_alignment:");
>
> @@ -870,7 +870,7 @@ vect_compute_data_ref_alignment (struct data_refer
>
>        if (dr_step % GET_MODE_SIZE (TYPE_MODE (vectype)) == 0)
>          {
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_NOTE, vect_location,
>                               "inner step divides the vector-size.");
>    misalign = STMT_VINFO_DR_INIT (stmt_info);
> @@ -879,7 +879,7 @@ vect_compute_data_ref_alignment (struct data_refer
>          }
>        else
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "inner step doesn't divide the vector-size.");
>    misalign = NULL_TREE;
> @@ -898,7 +898,7 @@ vect_compute_data_ref_alignment (struct data_refer
>
>        if (dr_step % GET_MODE_SIZE (TYPE_MODE (vectype)) != 0)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "SLP: step doesn't divide the vector-size.");
>    misalign = NULL_TREE;
> @@ -911,7 +911,7 @@ vect_compute_data_ref_alignment (struct data_refer
>    if ((aligned_to && tree_int_cst_compare (aligned_to, alignment) < 0)
>        || !misalign)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                             "Unknown alignment for access: ");
> @@ -941,7 +941,7 @@ vect_compute_data_ref_alignment (struct data_refer
>        if (!vect_can_force_dr_alignment_p (base, TYPE_ALIGN (vectype))
>    || (TREE_STATIC (base) && flag_section_anchors))
>   {
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                                 "can't force alignment of ref: ");
> @@ -953,7 +953,7 @@ vect_compute_data_ref_alignment (struct data_refer
>        /* Force the alignment of the decl.
>   NOTE: This is the only change to the code we make during
>   the analysis phase, before deciding to vectorize the loop.  */
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_NOTE, vect_location, "force alignment of ");
>            dump_generic_expr (MSG_NOTE, TDF_SLIM, ref);
> @@ -987,7 +987,7 @@ vect_compute_data_ref_alignment (struct data_refer
>    if (!host_integerp (misalign, 1))
>      {
>        /* Negative or overflowed misalignment value.  */
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "unexpected misalign value");
>        return false;
> @@ -995,7 +995,7 @@ vect_compute_data_ref_alignment (struct data_refer
>
>    SET_DR_MISALIGNMENT (dr, TREE_INT_CST_LOW (misalign));
>
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                         "misalign = %d bytes of ref ", DR_MISALIGNMENT (dr));
> @@ -1095,7 +1095,7 @@ vect_update_misalignment_for_peel (struct data_ref
>        return;
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "Setting misalignment to -1.");
>    SET_DR_MISALIGNMENT (dr, -1);
>  }
> @@ -1142,7 +1142,7 @@ vect_verify_datarefs_alignment (loop_vec_info loop
>        supportable_dr_alignment = vect_supportable_dr_alignment (dr, false);
>        if (!supportable_dr_alignment)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                if (DR_IS_READ (dr))
>                  dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
> @@ -1157,8 +1157,7 @@ vect_verify_datarefs_alignment (loop_vec_info loop
>              }
>            return false;
>          }
> -      if (supportable_dr_alignment != dr_aligned
> -          && dump_kind_p (MSG_NOTE))
> +      if (supportable_dr_alignment != dr_aligned && dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>                           "Vectorizing an unaligned access.");
>      }
> @@ -1215,7 +1214,7 @@ vector_alignment_reachable_p (struct data_referenc
>      {
>        HOST_WIDE_INT elmsize =
>   int_cst_value (TYPE_SIZE_UNIT (TREE_TYPE (vectype)));
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_NOTE, vect_location,
>                             "data size =" HOST_WIDE_INT_PRINT_DEC, elmsize);
> @@ -1224,7 +1223,7 @@ vector_alignment_reachable_p (struct data_referenc
>   }
>        if (DR_MISALIGNMENT (dr) % elmsize)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "data size does not divide the misalignment.\n");
>    return false;
> @@ -1235,7 +1234,7 @@ vector_alignment_reachable_p (struct data_referenc
>      {
>        tree type = TREE_TYPE (DR_REF (dr));
>        bool is_packed = not_size_aligned (DR_REF (dr));
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "Unknown misalignment, is_packed = %d",is_packed);
>        if (targetm.vectorize.vector_alignment_reachable (type, is_packed))
> @@ -1269,7 +1268,7 @@ vect_get_data_access_cost (struct data_reference *
>    else
>      vect_get_store_cost (dr, ncopies, inside_cost, body_cost_vec);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "vect_get_data_access_cost: inside_cost = %d, "
>                       "outside_cost = %d.", *inside_cost, *outside_cost);
> @@ -1567,7 +1566,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo
>    unsigned int nelements, mis, same_align_drs_max = 0;
>    stmt_vector_for_cost body_cost_vec = NULL;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "=== vect_enhance_data_refs_alignment ===");
>
> @@ -1622,7 +1621,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo
>   and so we can't generate the new base for the pointer.  */
>        if (STMT_VINFO_STRIDE_LOAD_P (stmt_info))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "strided load prevents peeling");
>    do_peeling = false;
> @@ -1738,7 +1737,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo
>          {
>            if (!aligned_access_p (dr))
>              {
> -              if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +              if (dump_enabled_p ())
>                  dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                   "vector alignment may not be reachable");
>                break;
> @@ -1879,7 +1878,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo
>    if (STMT_VINFO_GROUPED_ACCESS (stmt_info))
>      npeel /= GROUP_SIZE (stmt_info);
>
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_NOTE, vect_location,
>                               "Try peeling by %d", npeel);
>          }
> @@ -1951,7 +1950,7 @@ vect_enhance_data_refs_alignment (loop_vec_info lo
>            else
>              LOOP_PEELING_FOR_ALIGNMENT (loop_vinfo) = DR_MISALIGNMENT (dr0);
>    SET_DR_MISALIGNMENT (dr0, 0);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_NOTE, vect_location,
>                                 "Alignment of access forced using peeling.");
> @@ -2077,12 +2076,12 @@ vect_enhance_data_refs_alignment (loop_vec_info lo
>            stmt_vec_info stmt_info = vinfo_for_stmt (stmt);
>            dr = STMT_VINFO_DATA_REF (stmt_info);
>    SET_DR_MISALIGNMENT (dr, 0);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>              dump_printf_loc (MSG_NOTE, vect_location,
>                               "Alignment of access forced using versioning.");
>          }
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>                           "Versioning for alignment will be applied.");
>
> @@ -2148,7 +2147,7 @@ vect_find_same_alignment_drs (struct data_dependen
>      {
>        int dist = dist_v[loop_depth];
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>                           "dependence distance  = %d.", dist);
>
> @@ -2159,7 +2158,7 @@ vect_find_same_alignment_drs (struct data_dependen
>    /* Two references with distance zero have the same alignment.  */
>    VEC_safe_push (dr_p, heap, STMT_VINFO_SAME_ALIGN_REFS (stmtinfo_a), drb);
>    VEC_safe_push (dr_p, heap, STMT_VINFO_SAME_ALIGN_REFS (stmtinfo_b), dra);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>                dump_printf_loc (MSG_NOTE, vect_location,
>                                 "accesses have the same alignment.");
> @@ -2183,7 +2182,7 @@ bool
>  vect_analyze_data_refs_alignment (loop_vec_info loop_vinfo,
>                                    bb_vec_info bb_vinfo)
>  {
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "=== vect_analyze_data_refs_alignment ===");
>
> @@ -2201,7 +2200,7 @@ vect_analyze_data_refs_alignment (loop_vec_info lo
>
>    if (!vect_compute_data_refs_alignment (loop_vinfo, bb_vinfo))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "not vectorized: can't calculate alignment "
>                           "for data ref.");
> @@ -2254,7 +2253,7 @@ vect_analyze_group_access (struct data_reference *
>   {
>    GROUP_FIRST_ELEMENT (vinfo_for_stmt (stmt)) = stmt;
>    GROUP_SIZE (vinfo_for_stmt (stmt)) = groupsize;
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                                 "Detected single element interleaving ");
> @@ -2265,13 +2264,13 @@ vect_analyze_group_access (struct data_reference *
>
>    if (loop_vinfo)
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>                                   "Data access with gaps requires scalar "
>                                   "epilogue loop");
>                if (loop->inner)
>                  {
> -                  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                  if (dump_enabled_p ())
>                      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                       "Peeling for outer loop is not"
>                                       " supported");
> @@ -2284,7 +2283,7 @@ vect_analyze_group_access (struct data_reference *
>    return true;
>   }
>
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                             "not consecutive access ");
> @@ -2324,7 +2323,7 @@ vect_analyze_group_access (struct data_reference *
>              {
>                if (DR_IS_WRITE (data_ref))
>                  {
> -                  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                  if (dump_enabled_p ())
>                      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                       "Two store stmts share the same dr.");
>                    return false;
> @@ -2335,7 +2334,7 @@ vect_analyze_group_access (struct data_reference *
>                if (GROUP_READ_WRITE_DEPENDENCE (vinfo_for_stmt (next))
>                    || GROUP_READ_WRITE_DEPENDENCE (vinfo_for_stmt (prev)))
>                  {
> -                  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                  if (dump_enabled_p ())
>                      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                       "READ_WRITE dependence in interleaving.");
>                    return false;
> @@ -2355,7 +2354,7 @@ vect_analyze_group_access (struct data_reference *
>            next_step = DR_STEP (STMT_VINFO_DATA_REF (vinfo_for_stmt (next)));
>            if (tree_int_cst_compare (step, next_step))
>              {
> -              if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +              if (dump_enabled_p ())
>                  dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                   "not consecutive access in interleaving");
>                return false;
> @@ -2372,7 +2371,7 @@ vect_analyze_group_access (struct data_reference *
>        slp_impossible = true;
>        if (DR_IS_WRITE (data_ref))
>   {
> -                  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                  if (dump_enabled_p ())
>                      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                       "interleaved store with gaps");
>    return false;
> @@ -2401,7 +2400,7 @@ vect_analyze_group_access (struct data_reference *
>           greater than STEP.  */
>        if (dr_step && dr_step < count_in_bytes + gaps * type_size)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                 "interleaving size is greater than step for ");
> @@ -2424,7 +2423,7 @@ vect_analyze_group_access (struct data_reference *
>              }
>            else
>              {
> -              if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +              if (dump_enabled_p ())
>                  dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                   "interleaved store with gaps");
>                return false;
> @@ -2434,7 +2433,7 @@ vect_analyze_group_access (struct data_reference *
>        /* Check that STEP is a multiple of type size.  */
>        if (dr_step && (dr_step % type_size) != 0)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                 "step is not a multiple of type size: step ");
> @@ -2450,7 +2449,7 @@ vect_analyze_group_access (struct data_reference *
>          groupsize = count;
>
>        GROUP_SIZE (vinfo_for_stmt (stmt)) = groupsize;
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>                           "Detected interleaving of size %d", (int)groupsize);
>
> @@ -2469,13 +2468,13 @@ vect_analyze_group_access (struct data_reference *
>        /* There is a gap in the end of the group.  */
>        if (groupsize - last_accessed_element > 0 && loop_vinfo)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "Data access with gaps requires scalar "
>                               "epilogue loop");
>            if (loop->inner)
>              {
> -              if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +              if (dump_enabled_p ())
>                  dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                   "Peeling for outer loop is not supported");
>                return false;
> @@ -2508,7 +2507,7 @@ vect_analyze_data_ref_access (struct data_referenc
>
>    if (loop_vinfo && !step)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "bad data-ref access in loop");
>        return false;
> @@ -2531,7 +2530,7 @@ vect_analyze_data_ref_access (struct data_referenc
>        step = STMT_VINFO_DR_STEP (stmt_info);
>        if (integer_zerop (step))
>   {
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                               "zero step in outer loop.");
>    if (DR_IS_READ (dr))
> @@ -2557,7 +2556,7 @@ vect_analyze_data_ref_access (struct data_referenc
>
>    if (loop && nested_in_vect_loop_p (loop, stmt))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>                           "grouped access in outer loop.");
>        return false;
> @@ -2588,7 +2587,7 @@ vect_analyze_data_ref_accesses (loop_vec_info loop
>    VEC (data_reference_p, heap) *datarefs;
>    struct data_reference *dr;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "=== vect_analyze_data_ref_accesses ===");
>
> @@ -2601,7 +2600,7 @@ vect_analyze_data_ref_accesses (loop_vec_info loop
>      if (STMT_VINFO_VECTORIZABLE (vinfo_for_stmt (DR_STMT (dr)))
>          && !vect_analyze_data_ref_access (dr))
>        {
> - if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> + if (dump_enabled_p ())
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                             "not vectorized: complicated access pattern.");
>
> @@ -2631,7 +2630,7 @@ vect_prune_runtime_alias_test_list (loop_vec_info
>      LOOP_VINFO_MAY_ALIAS_DDRS (loop_vinfo);
>    unsigned i, j;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "=== vect_prune_runtime_alias_test_list ===");
>
> @@ -2649,7 +2648,7 @@ vect_prune_runtime_alias_test_list (loop_vec_info
>
>    if (vect_vfa_range_equal (ddr_i, ddr_j))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_NOTE, vect_location,
>                                     "found equal ranges ");
> @@ -2677,7 +2676,7 @@ vect_prune_runtime_alias_test_list (loop_vec_info
>    if (VEC_length (ddr_p, ddrs) >
>         (unsigned) PARAM_VALUE (PARAM_VECT_MAX_VERSION_FOR_ALIAS_CHECKS))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION,  vect_location,
>                             "disable versioning for alias - max number of "
> @@ -2964,7 +2963,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>    tree scalar_type;
>    bool res, stop_bb_analysis = false;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "=== vect_analyze_data_refs ===\n");
>
> @@ -2979,7 +2978,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>
>        if (!res)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "not vectorized: loop contains function calls"
>                               " or data references that cannot be analyzed");
> @@ -3011,7 +3010,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>        if (!compute_all_dependences (BB_VINFO_DATAREFS (bb_vinfo),
>      &BB_VINFO_DDRS (bb_vinfo), NULL, true))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "not vectorized: basic block contains function"
>                               " calls or data references that cannot be"
> @@ -3035,7 +3034,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>
>        if (!dr || !DR_REF (dr))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "not vectorized: unhandled data-ref ");
>            return false;
> @@ -3081,7 +3080,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>
>    if (!gather)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                     "not vectorized: data ref analysis "
> @@ -3102,7 +3101,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>
>        if (TREE_CODE (DR_BASE_ADDRESS (dr)) == INTEGER_CST)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "not vectorized: base addr of dr is a "
>                               "constant");
> @@ -3121,7 +3120,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>
>        if (TREE_THIS_VOLATILE (DR_REF (dr)))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                 "not vectorized: volatile type ");
> @@ -3140,7 +3139,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>
>        if (stmt_can_throw_internal (stmt))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                 "not vectorized: statement can throw an "
> @@ -3163,7 +3162,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>        if (TREE_CODE (DR_REF (dr)) == COMPONENT_REF
>    && DECL_BIT_FIELD (TREE_OPERAND (DR_REF (dr), 1)))
>   {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                 "not vectorized: statement is bitfield "
> @@ -3189,7 +3188,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>
>        if (is_gimple_call (stmt))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_MISSED_OPTIMIZATION,  vect_location,
>                                 "not vectorized: dr in a call ");
> @@ -3232,7 +3231,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>            tree inner_base = build_fold_indirect_ref
>                                  (fold_build_pointer_plus (base, init));
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                                 "analyze in outer-loop: ");
> @@ -3245,7 +3244,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>
>    if (pbitpos % BITS_PER_UNIT != 0)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                   "failed: bit offset alignment.\n");
>        return false;
> @@ -3255,7 +3254,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>    if (!simple_iv (loop, loop_containing_stmt (stmt), outer_base,
>                            &base_iv, false))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                   "failed: evolution of base is not affine.\n");
>        return false;
> @@ -3278,7 +3277,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>    else if (!simple_iv (loop, loop_containing_stmt (stmt), poffset,
>                                 &offset_iv, false))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                   "evolution of offset is not affine.\n");
>        return false;
> @@ -3303,7 +3302,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>    STMT_VINFO_DR_ALIGNED_TO (stmt_info) =
>   size_int (highest_pow2_factor (offset_iv.base));
>
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                                 "\touter base_address: ");
> @@ -3327,7 +3326,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>
>        if (STMT_VINFO_DATA_REF (stmt_info))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                 "not vectorized: more than one data ref "
> @@ -3355,7 +3354,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>                  get_vectype_for_scalar_type (scalar_type);
>        if (!STMT_VINFO_VECTYPE (stmt_info))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                 "not vectorized: no vectype for stmt: ");
> @@ -3406,7 +3405,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>      {
>        STMT_VINFO_DATA_REF (stmt_info) = NULL;
>        free_data_ref (dr);
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                     "not vectorized: not suitable for gather "
> @@ -3459,7 +3458,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>
>    if (bad)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                     "not vectorized: data dependence conflict"
> @@ -3480,7 +3479,7 @@ vect_analyze_data_refs (loop_vec_info loop_vinfo,
>        = vect_check_strided_load (stmt, loop_vinfo, NULL, NULL);
>    if (!strided_load)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                     "not vectorized: not suitable for strided "
> @@ -3668,7 +3667,7 @@ vect_create_addr_base_for_vector_ref (gimple stmt,
>   mark_ptr_info_alignment_unknown (SSA_NAME_PTR_INFO (vec_stmt));
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "created ");
>        dump_generic_expr (MSG_NOTE, TDF_SLIM, vec_stmt);
> @@ -3790,7 +3789,7 @@ vect_create_data_ref_ptr (gimple stmt, tree aggr_t
>       in LOOP.  */
>    base_name = build_fold_indirect_ref (unshare_expr (DR_BASE_ADDRESS (dr)));
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        tree data_ref_base = base_name;
>        dump_printf_loc (MSG_NOTE, vect_location,
> @@ -4120,7 +4119,7 @@ vect_grouped_store_supported (tree vectype, unsign
>    /* vect_permute_store_chain requires the group size to be a power of two.  */
>    if (exact_log2 (count) == -1)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "the size of the group of accesses"
>                           " is not a power of 2");
> @@ -4146,7 +4145,7 @@ vect_grouped_store_supported (tree vectype, unsign
>   }
>      }
>
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf (MSG_MISSED_OPTIMIZATION,
>                   "interleave op not supported by target.");
>    return false;
> @@ -4564,7 +4563,7 @@ vect_grouped_load_supported (tree vectype, unsigne
>    /* vect_permute_load_chain requires the group size to be a power of two.  */
>    if (exact_log2 (count) == -1)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "the size of the group of accesses"
>                           " is not a power of 2");
> @@ -4588,7 +4587,7 @@ vect_grouped_load_supported (tree vectype, unsigne
>   }
>      }
>
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                       "extract even/odd not supported by target");
>    return false;
> Index: tree-vect-patterns.c
> ===================================================================
> --- tree-vect-patterns.c (revision 192695)
> +++ tree-vect-patterns.c (working copy)
> @@ -416,7 +416,7 @@ vect_recog_dot_prod_pattern (VEC (gimple, heap) **
>    pattern_stmt = gimple_build_assign_with_ops (DOT_PROD_EXPR, var,
>         oprnd00, oprnd01, oprnd1);
>
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                         "vect_recog_dot_prod_pattern: detected: ");
> @@ -676,7 +676,7 @@ vect_recog_widen_mult_pattern (VEC (gimple, heap)
>      return NULL;
>
>    /* Pattern detected.  */
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                       "vect_recog_widen_mult_pattern: detected: ");
>
> @@ -699,7 +699,7 @@ vect_recog_widen_mult_pattern (VEC (gimple, heap)
>    pattern_stmt = gimple_build_assign_with_ops (WIDEN_MULT_EXPR, var, oprnd0,
>         oprnd1);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_gimple_stmt_loc (MSG_NOTE, vect_location, TDF_SLIM, pattern_stmt, 0);
>
>    VEC_safe_push (gimple, heap, *stmts, last_stmt);
> @@ -912,7 +912,7 @@ vect_recog_widen_sum_pattern (VEC (gimple, heap) *
>    pattern_stmt = gimple_build_assign_with_ops (WIDEN_SUM_EXPR, var,
>         oprnd0, oprnd1);
>
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                         "vect_recog_widen_sum_pattern: detected: ");
> @@ -1217,7 +1217,7 @@ vect_recog_over_widening_pattern (VEC (gimple, hea
>        STMT_VINFO_RELATED_STMT (vinfo_for_stmt (stmt)) = pattern_stmt;
>        new_pattern_def_seq (vinfo_for_stmt (stmt), new_def_stmt);
>
> -      if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                             "created pattern stmt: ");
> @@ -1285,7 +1285,7 @@ vect_recog_over_widening_pattern (VEC (gimple, hea
>      return NULL;
>
>    /* Pattern detected.  */
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                         "vect_recog_over_widening_pattern: detected: ");
> @@ -1421,7 +1421,7 @@ vect_recog_widen_shift_pattern (VEC (gimple, heap)
>      return NULL;
>
>    /* Pattern detected.  */
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                       "vect_recog_widen_shift_pattern: detected: ");
>
> @@ -1445,7 +1445,7 @@ vect_recog_widen_shift_pattern (VEC (gimple, heap)
>    pattern_stmt =
>      gimple_build_assign_with_ops (WIDEN_LSHIFT_EXPR, var, oprnd0, oprnd1);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_gimple_stmt_loc (MSG_NOTE, vect_location, TDF_SLIM, pattern_stmt, 0);
>
>    VEC_safe_push (gimple, heap, *stmts, last_stmt);
> @@ -1567,7 +1567,7 @@ vect_recog_vector_vector_shift_pattern (VEC (gimpl
>      }
>
>    /* Pattern detected.  */
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                       "vect_recog_vector_vector_shift_pattern: detected: ");
>
> @@ -1575,7 +1575,7 @@ vect_recog_vector_vector_shift_pattern (VEC (gimpl
>    var = vect_recog_temp_ssa_var (TREE_TYPE (oprnd0), NULL);
>    pattern_stmt = gimple_build_assign_with_ops (rhs_code, var, oprnd0, def);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_gimple_stmt_loc (MSG_NOTE, vect_location, TDF_SLIM, pattern_stmt, 0);
>
>    VEC_safe_push (gimple, heap, *stmts, last_stmt);
> @@ -1685,7 +1685,7 @@ vect_recog_divmod_pattern (VEC (gimple, heap) **st
>   return NULL;
>
>        /* Pattern detected.  */
> -      if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                           "vect_recog_divmod_pattern: detected: ");
>
> @@ -1789,7 +1789,7 @@ vect_recog_divmod_pattern (VEC (gimple, heap) **st
>      signmask);
>   }
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_gimple_stmt_loc (MSG_NOTE, vect_location, TDF_SLIM, pattern_stmt,
>                                0);
>
> @@ -2031,7 +2031,7 @@ vect_recog_divmod_pattern (VEC (gimple, heap) **st
>      }
>
>    /* Pattern detected.  */
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                         "vect_recog_divmod_pattern: detected: ");
> @@ -2199,7 +2199,7 @@ vect_recog_mixed_size_cond_pattern (VEC (gimple, h
>    *type_in = vecitype;
>    *type_out = vectype;
>
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                       "vect_recog_mixed_size_cond_pattern: detected: ");
>
> @@ -2592,7 +2592,7 @@ vect_recog_bool_pattern (VEC (gimple, heap) **stmt
>        *type_out = vectype;
>        *type_in = vectype;
>        VEC_safe_push (gimple, heap, *stmts, last_stmt);
> -      if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                           "vect_recog_bool_pattern: detected: ");
>
> @@ -2638,7 +2638,7 @@ vect_recog_bool_pattern (VEC (gimple, heap) **stmt
>        *type_out = vectype;
>        *type_in = vectype;
>        VEC_safe_push (gimple, heap, *stmts, last_stmt);
> -      if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                           "vect_recog_bool_pattern: detected: ");
>        return pattern_stmt;
> @@ -2788,7 +2788,7 @@ vect_pattern_recog_1 (vect_recog_func_ptr vect_rec
>      }
>
>    /* Found a vectorizable pattern.  */
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                         "pattern recognized: ");
> @@ -2814,7 +2814,7 @@ vect_pattern_recog_1 (vect_recog_func_ptr vect_rec
>      {
>        stmt_info = vinfo_for_stmt (stmt);
>        pattern_stmt = STMT_VINFO_RELATED_STMT (stmt_info);
> -      if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>                             "additional pattern stmt: ");
> @@ -2915,7 +2915,7 @@ vect_pattern_recog (loop_vec_info loop_vinfo, bb_v
>    VEC (gimple, heap) *stmts_to_replace = VEC_alloc (gimple, heap, 1);
>    gimple stmt;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "=== vect_pattern_recog ===");
>
> Index: tree-vect-stmts.c
> ===================================================================
> --- tree-vect-stmts.c (revision 192695)
> +++ tree-vect-stmts.c (working copy)
> @@ -190,7 +190,7 @@ vect_mark_relevant (VEC(gimple,heap) **worklist, g
>    bool save_live_p = STMT_VINFO_LIVE_P (stmt_info);
>    gimple pattern_stmt;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "mark relevant %d, live %d.", relevant, live_p);
>
> @@ -246,7 +246,7 @@ vect_mark_relevant (VEC(gimple,heap) **worklist, g
>
>            pattern_stmt = STMT_VINFO_RELATED_STMT (stmt_info);
>
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_NOTE, vect_location,
>                               "last stmt in pattern. don't mark"
>                               " relevant/live.");
> @@ -265,7 +265,7 @@ vect_mark_relevant (VEC(gimple,heap) **worklist, g
>    if (STMT_VINFO_RELEVANT (stmt_info) == save_relevant
>        && STMT_VINFO_LIVE_P (stmt_info) == save_live_p)
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>                           "already marked relevant/live.");
>        return;
> @@ -310,7 +310,7 @@ vect_stmt_relevant_p (gimple stmt, loop_vec_info l
>    if (gimple_code (stmt) != GIMPLE_PHI)
>      if (gimple_vdef (stmt))
>        {
> - if (dump_kind_p (MSG_NOTE))
> + if (dump_enabled_p ())
>    dump_printf_loc (MSG_NOTE, vect_location,
>                             "vec_stmt_relevant_p: stmt has vdefs.");
>   *relevant = vect_used_in_scope;
> @@ -324,7 +324,7 @@ vect_stmt_relevant_p (gimple stmt, loop_vec_info l
>    basic_block bb = gimple_bb (USE_STMT (use_p));
>    if (!flow_bb_inside_loop_p (loop, bb))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>                                   "vec_stmt_relevant_p: used out of loop.");
>
> @@ -437,7 +437,7 @@ process_use (gimple stmt, tree use, loop_vec_info
>
>    if (!vect_is_simple_use (use, stmt, loop_vinfo, NULL, &def_stmt, &def, &dt))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "not vectorized: unsupported use in stmt.");
>        return false;
> @@ -449,7 +449,7 @@ process_use (gimple stmt, tree use, loop_vec_info
>    def_bb = gimple_bb (def_stmt);
>    if (!flow_bb_inside_loop_p (loop, def_bb))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location, "def_stmt is out of loop.");
>        return true;
>      }
> @@ -467,7 +467,7 @@ process_use (gimple stmt, tree use, loop_vec_info
>        && STMT_VINFO_DEF_TYPE (dstmt_vinfo) == vect_reduction_def
>        && bb->loop_father == def_bb->loop_father)
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>                           "reduc-stmt defining reduc-phi in the same nest.");
>        if (STMT_VINFO_IN_PATTERN_P (dstmt_vinfo))
> @@ -487,7 +487,7 @@ process_use (gimple stmt, tree use, loop_vec_info
>   ...  */
>    if (flow_loop_nested_p (def_bb->loop_father, bb->loop_father))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>                           "outer-loop def-stmt defining inner-loop stmt.");
>
> @@ -525,7 +525,7 @@ process_use (gimple stmt, tree use, loop_vec_info
>   stmt # use (d) */
>    else if (flow_loop_nested_p (bb->loop_father, def_bb->loop_father))
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>                           "inner-loop def-stmt defining outer-loop stmt.");
>
> @@ -589,7 +589,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo
>    enum vect_relevant relevant, tmp_relevant;
>    enum vect_def_type def_type;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "=== vect_mark_stmts_to_be_vectorized ===");
>
> @@ -602,7 +602,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo
>        for (si = gsi_start_phis (bb); !gsi_end_p (si); gsi_next (&si))
>   {
>    phi = gsi_stmt (si);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "init: phi relevant? ");
>        dump_gimple_stmt (MSG_NOTE, TDF_SLIM, phi, 0);
> @@ -614,7 +614,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo
>        for (si = gsi_start_bb (bb); !gsi_end_p (si); gsi_next (&si))
>   {
>    stmt = gsi_stmt (si);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "init: stmt relevant? ");
>        dump_gimple_stmt (MSG_NOTE, TDF_SLIM, stmt, 0);
> @@ -632,7 +632,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo
>        ssa_op_iter iter;
>
>        stmt = VEC_pop (gimple, worklist);
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>            dump_printf_loc (MSG_NOTE, vect_location, "worklist:
> examine stmt: ");
>            dump_gimple_stmt (MSG_NOTE, TDF_SLIM, stmt, 0);
> @@ -677,7 +677,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo
>              /* fall through */
>
>          default:
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                       "unsupported use of reduction.");
>              VEC_free (gimple, heap, worklist);
> @@ -692,7 +692,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo
>                  && tmp_relevant != vect_used_in_outer_by_reduction
>                  && tmp_relevant != vect_used_in_outer)
>                {
> -                if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                if (dump_enabled_p ())
>                    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                     "unsupported use of nested cycle.");
>
> @@ -707,7 +707,7 @@ vect_mark_stmts_to_be_vectorized (loop_vec_info lo
>              if (tmp_relevant != vect_unused_in_scope
>                  && tmp_relevant != vect_used_by_reduction)
>                {
> -                if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                if (dump_enabled_p ())
>                    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                     "unsupported use of double reduction.");
>
> @@ -830,7 +830,7 @@ vect_model_simple_cost (stmt_vec_info stmt_info, i
>    inside_cost = record_stmt_cost (body_cost_vec, ncopies, vector_stmt,
>    stmt_info, 0, vect_body);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "vect_model_simple_cost: inside_cost = %d, "
>                       "prologue_cost = %d .", inside_cost, prologue_cost);
> @@ -876,7 +876,7 @@ vect_model_promotion_demotion_cost (stmt_vec_info
>        prologue_cost += add_stmt_cost (target_cost_data, 1, vector_stmt,
>        stmt_info, 0, vect_prologue);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "vect_model_promotion_demotion_cost: inside_cost = %d, "
>                       "prologue_cost = %d .", inside_cost, prologue_cost);
> @@ -960,7 +960,7 @@ vect_model_store_cost (stmt_vec_info stmt_info, in
>        inside_cost = record_stmt_cost (body_cost_vec, nstmts, vec_perm,
>        stmt_info, 0, vect_body);
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>                           "vect_model_store_cost: strided group_size = %d .",
>                           group_size);
> @@ -969,7 +969,7 @@ vect_model_store_cost (stmt_vec_info stmt_info, in
>    /* Costs of the stores.  */
>    vect_get_store_cost (first_dr, ncopies, &inside_cost, body_cost_vec);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "vect_model_store_cost: inside_cost = %d, "
>                       "prologue_cost = %d .", inside_cost, prologue_cost);
> @@ -994,7 +994,7 @@ vect_get_store_cost (struct data_reference *dr, in
>    vector_store, stmt_info, 0,
>    vect_body);
>
> -        if (dump_kind_p (MSG_NOTE))
> +        if (dump_enabled_p ())
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "vect_model_store_cost: aligned.");
>          break;
> @@ -1006,7 +1006,7 @@ vect_get_store_cost (struct data_reference *dr, in
>   *inside_cost += record_stmt_cost (body_cost_vec, ncopies,
>    unaligned_store, stmt_info,
>    DR_MISALIGNMENT (dr), vect_body);
> -        if (dump_kind_p (MSG_NOTE))
> +        if (dump_enabled_p ())
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "vect_model_store_cost: unaligned supported by "
>                             "hardware.");
> @@ -1017,7 +1017,7 @@ vect_get_store_cost (struct data_reference *dr, in
>        {
>          *inside_cost = VECT_MAX_COST;
>
> -        if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +        if (dump_enabled_p ())
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                             "vect_model_store_cost: unsupported access.");
>          break;
> @@ -1076,7 +1076,7 @@ vect_model_load_cost (stmt_vec_info stmt_info, int
>        inside_cost += record_stmt_cost (body_cost_vec, nstmts, vec_perm,
>         stmt_info, 0, vect_body);
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>                           "vect_model_load_cost: strided group_size = %d .",
>                           group_size);
> @@ -1100,7 +1100,7 @@ vect_model_load_cost (stmt_vec_info stmt_info, int
>   &inside_cost, &prologue_cost,
>   prologue_cost_vec, body_cost_vec, true);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "vect_model_load_cost: inside_cost = %d, "
>                       "prologue_cost = %d .", inside_cost, prologue_cost);
> @@ -1127,7 +1127,7 @@ vect_get_load_cost (struct data_reference *dr, int
>   *inside_cost += record_stmt_cost (body_cost_vec, ncopies, vector_load,
>    stmt_info, 0, vect_body);
>
> -        if (dump_kind_p (MSG_NOTE))
> +        if (dump_enabled_p ())
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "vect_model_load_cost: aligned.");
>
> @@ -1140,7 +1140,7 @@ vect_get_load_cost (struct data_reference *dr, int
>    unaligned_load, stmt_info,
>    DR_MISALIGNMENT (dr), vect_body);
>
> -        if (dump_kind_p (MSG_NOTE))
> +        if (dump_enabled_p ())
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "vect_model_load_cost: unaligned supported by "
>                             "hardware.");
> @@ -1161,7 +1161,7 @@ vect_get_load_cost (struct data_reference *dr, int
>    *inside_cost += record_stmt_cost (body_cost_vec, 1, vector_stmt,
>      stmt_info, 0, vect_body);
>
> -        if (dump_kind_p (MSG_NOTE))
> +        if (dump_enabled_p ())
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "vect_model_load_cost: explicit realign");
>
> @@ -1169,7 +1169,7 @@ vect_get_load_cost (struct data_reference *dr, int
>        }
>      case dr_explicit_realign_optimized:
>        {
> -        if (dump_kind_p (MSG_NOTE))
> +        if (dump_enabled_p ())
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "vect_model_load_cost: unaligned software "
>                             "pipelined.");
> @@ -1197,7 +1197,7 @@ vect_get_load_cost (struct data_reference *dr, int
>   *inside_cost += record_stmt_cost (body_cost_vec, ncopies, vec_perm,
>    stmt_info, 0, vect_body);
>
> -        if (dump_kind_p (MSG_NOTE))
> +        if (dump_enabled_p ())
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "vect_model_load_cost: explicit realign optimized");
>
> @@ -1208,7 +1208,7 @@ vect_get_load_cost (struct data_reference *dr, int
>        {
>          *inside_cost = VECT_MAX_COST;
>
> -        if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +        if (dump_enabled_p ())
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                             "vect_model_load_cost: unsupported access.");
>          break;
> @@ -1258,7 +1258,7 @@ vect_init_vector_1 (gimple stmt, gimple new_stmt,
>         }
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                         "created new init_stmt: ");
> @@ -1340,7 +1340,7 @@ vect_get_vec_def_for_operand (tree op, gimple stmt
>    bool is_simple_use;
>    tree vector_type;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                         "vect_get_vec_def_for_operand: ");
> @@ -1350,7 +1350,7 @@ vect_get_vec_def_for_operand (tree op, gimple stmt
>    is_simple_use = vect_is_simple_use (op, stmt, loop_vinfo, NULL,
>        &def_stmt, &def, &dt);
>    gcc_assert (is_simple_use);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        int loc_printed = 0;
>        if (def)
> @@ -1382,7 +1382,7 @@ vect_get_vec_def_for_operand (tree op, gimple stmt
>    *scalar_def = op;
>
>          /* Create 'vect_cst_ = {cst,cst,...,cst}'  */
> -        if (dump_kind_p (MSG_NOTE))
> +        if (dump_enabled_p ())
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "Create vector_cst. nunits = %d", nunits);
>
> @@ -1399,7 +1399,7 @@ vect_get_vec_def_for_operand (tree op, gimple stmt
>    *scalar_def = def;
>
>          /* Create 'vec_inv = {inv,inv,..,inv}'  */
> -        if (dump_kind_p (MSG_NOTE))
> +        if (dump_enabled_p ())
>            dump_printf_loc (MSG_NOTE, vect_location, "Create vector_inv.");
>
>          return vect_init_vector (stmt, def, vector_type, NULL);
> @@ -1661,7 +1661,7 @@ vect_finish_stmt_generation (gimple stmt, gimple v
>    set_vinfo_for_stmt (vec_stmt, new_stmt_vec_info (vec_stmt, loop_vinfo,
>                                                     bb_vinfo));
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "add new stmt: ");
>        dump_gimple_stmt (MSG_NOTE, TDF_SLIM, vec_stmt, 0);
> @@ -1764,7 +1764,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat
>        if (rhs_type
>    && !types_compatible_p (rhs_type, TREE_TYPE (op)))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "argument types differ.");
>    return false;
> @@ -1775,7 +1775,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat
>        if (!vect_is_simple_use_1 (op, stmt, loop_vinfo, bb_vinfo,
>   &def_stmt, &def, &dt[i], &opvectype))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "use not simple.");
>    return false;
> @@ -1786,7 +1786,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat
>        else if (opvectype
>         && opvectype != vectype_in)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "argument vector types differ.");
>    return false;
> @@ -1800,7 +1800,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat
>      gcc_assert (vectype_in);
>    if (!vectype_in)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                             "no vectype for scalar type ");
> @@ -1829,7 +1829,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat
>    fndecl = vectorizable_function (stmt, vectype_out, vectype_in);
>    if (fndecl == NULL_TREE)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "function is not vectorizable.");
>
> @@ -1852,7 +1852,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat
>    if (!vec_stmt) /* transformation not required.  */
>      {
>        STMT_VINFO_TYPE (stmt_info) = call_vec_info_type;
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location, "=== vectorizable_call ===");
>        vect_model_simple_cost (stmt_info, ncopies, dt, NULL, NULL);
>        return true;
> @@ -1860,7 +1860,7 @@ vectorizable_call (gimple stmt, gimple_stmt_iterat
>
>    /** Transform.  **/
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "transform call.");
>
>    /* Handle def.  */
> @@ -2375,7 +2375,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_
>    && (TYPE_PRECISION (rhs_type)
>        != GET_MODE_PRECISION (TYPE_MODE (rhs_type)))))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "type conversion to/from bit-precision unsupported.");
>        return false;
> @@ -2385,7 +2385,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_
>    if (!vect_is_simple_use_1 (op0, stmt, loop_vinfo, bb_vinfo,
>       &def_stmt, &def, &dt[0], &vectype_in))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "use not simple.");
>        return false;
> @@ -2407,7 +2407,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_
>
>        if (!ok)
>   {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "use not simple.");
>    return false;
> @@ -2422,7 +2422,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_
>      gcc_assert (vectype_in);
>    if (!vectype_in)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                             "no vectype for scalar type ");
> @@ -2466,7 +2466,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_
>   break;
>        /* FALLTHRU */
>      unsupported:
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "conversion not supported by target.");
>        return false;
> @@ -2565,7 +2565,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_
>
>    if (!vec_stmt) /* transformation not required.  */
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>                           "=== vectorizable_conversion ===");
>        if (code == FIX_TRUNC_EXPR || code == FLOAT_EXPR)
> @@ -2588,7 +2588,7 @@ vectorizable_conversion (gimple stmt, gimple_stmt_
>      }
>
>    /** Transform.  **/
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "transform conversion. ncopies = %d.", ncopies);
>
> @@ -2941,7 +2941,7 @@ vectorizable_assignment (gimple stmt, gimple_stmt_
>    if (!vect_is_simple_use_1 (op, stmt, loop_vinfo, bb_vinfo,
>       &def_stmt, &def, &dt[0], &vectype_in))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "use not simple.");
>        return false;
> @@ -2970,7 +2970,7 @@ vectorizable_assignment (gimple stmt, gimple_stmt_
>      > TYPE_PRECISION (TREE_TYPE (op)))
>     && TYPE_UNSIGNED (TREE_TYPE (op))))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "type conversion to/from bit-precision "
>                           "unsupported.");
> @@ -2980,7 +2980,7 @@ vectorizable_assignment (gimple stmt, gimple_stmt_
>    if (!vec_stmt) /* transformation not required.  */
>      {
>        STMT_VINFO_TYPE (stmt_info) = assignment_vec_info_type;
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>                           "=== vectorizable_assignment ===");
>        vect_model_simple_cost (stmt_info, ncopies, dt, NULL, NULL);
> @@ -2988,7 +2988,7 @@ vectorizable_assignment (gimple stmt, gimple_stmt_
>      }
>
>    /** Transform.  **/
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "transform assignment.");
>
>    /* Handle def.  */
> @@ -3135,7 +3135,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>    if (TYPE_PRECISION (TREE_TYPE (scalar_dest))
>        != GET_MODE_PRECISION (TYPE_MODE (TREE_TYPE (scalar_dest))))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "bit-precision shifts not supported.");
>        return false;
> @@ -3145,7 +3145,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>    if (!vect_is_simple_use_1 (op0, stmt, loop_vinfo, bb_vinfo,
>                               &def_stmt, &def, &dt[0], &vectype))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "use not simple.");
>        return false;
> @@ -3158,7 +3158,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>      gcc_assert (vectype);
>    if (!vectype)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "no vectype for scalar type ");
>        return false;
> @@ -3173,7 +3173,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>    if (!vect_is_simple_use_1 (op1, stmt, loop_vinfo, bb_vinfo, &def_stmt,
>       &def, &dt[1], &op1_vectype))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "use not simple.");
>        return false;
> @@ -3218,7 +3218,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>      }
>    else
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "operand mode requires invariant argument.");
>        return false;
> @@ -3228,7 +3228,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>    if (!scalar_shift_arg)
>      {
>        optab = optab_for_tree_code (code, vectype, optab_vector);
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>                           "vector/vector shift/rotate found.");
>
> @@ -3237,7 +3237,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>        if (op1_vectype == NULL_TREE
>    || TYPE_MODE (op1_vectype) != TYPE_MODE (vectype))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "unusable type for last operand in"
>                               " vector/vector shift/rotate.");
> @@ -3252,7 +3252,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>        if (optab
>            && optab_handler (optab, TYPE_MODE (vectype)) != CODE_FOR_nothing)
>          {
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_NOTE, vect_location,
>                               "vector/scalar shift/rotate found.");
>          }
> @@ -3265,7 +3265,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>              {
>        scalar_shift_arg = false;
>
> -              if (dump_kind_p (MSG_NOTE))
> +              if (dump_enabled_p ())
>                  dump_printf_loc (MSG_NOTE, vect_location,
>                                   "vector/vector shift/rotate found.");
>
> @@ -3282,7 +3282,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>        && TYPE_MODE (TREE_TYPE (vectype))
>   != TYPE_MODE (TREE_TYPE (op1)))
>      {
> -                      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                      if (dump_enabled_p ())
>                          dump_printf_loc (MSG_MISSED_OPTIMIZATION,
> vect_location,
>                                           "unusable type for last operand in"
>                                           " vector/vector shift/rotate.");
> @@ -3302,7 +3302,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>    /* Supportable by target?  */
>    if (!optab)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "no optab.");
>        return false;
> @@ -3311,7 +3311,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>    icode = (int) optab_handler (optab, vec_mode);
>    if (icode == CODE_FOR_nothing)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "op not supported by target.");
>        /* Check only during analysis.  */
> @@ -3319,7 +3319,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>            || (vf < vect_min_worthwhile_factor (code)
>                && !vec_stmt))
>          return false;
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location, "proceeding using
> word mode.");
>      }
>
> @@ -3328,7 +3328,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>        && vf < vect_min_worthwhile_factor (code)
>        && !vec_stmt)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "not worthwhile without SIMD support.");
>        return false;
> @@ -3337,7 +3337,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>    if (!vec_stmt) /* transformation not required.  */
>      {
>        STMT_VINFO_TYPE (stmt_info) = shift_vec_info_type;
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location, "===
> vectorizable_shift ===");
>        vect_model_simple_cost (stmt_info, ncopies, dt, NULL, NULL);
>        return true;
> @@ -3345,7 +3345,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>
>    /** Transform.  **/
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "transform binary/unary operation.");
>
> @@ -3382,7 +3382,7 @@ vectorizable_shift (gimple stmt, gimple_stmt_itera
>                optab_op2_mode = insn_data[icode].operand[2].mode;
>                if (!VECTOR_MODE_P (optab_op2_mode))
>                  {
> -                  if (dump_kind_p (MSG_NOTE))
> +                  if (dump_enabled_p ())
>                      dump_printf_loc (MSG_NOTE, vect_location,
>                                       "operand 1 using scalar mode.");
>                    vec_oprnd1 = op1;
> @@ -3510,7 +3510,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i
>    op_type = TREE_CODE_LENGTH (code);
>    if (op_type != unary_op && op_type != binary_op && op_type != ternary_op)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "num. args = %d (not unary/binary/ternary op).",
>                           op_type);
> @@ -3529,7 +3529,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i
>        && code != BIT_XOR_EXPR
>        && code != BIT_AND_EXPR)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "bit-precision arithmetic not supported.");
>        return false;
> @@ -3539,7 +3539,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i
>    if (!vect_is_simple_use_1 (op0, stmt, loop_vinfo, bb_vinfo,
>       &def_stmt, &def, &dt[0], &vectype))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "use not simple.");
>        return false;
> @@ -3552,7 +3552,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i
>      gcc_assert (vectype);
>    if (!vectype)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                             "no vectype for scalar type ");
> @@ -3574,7 +3574,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i
>        if (!vect_is_simple_use (op1, stmt, loop_vinfo, bb_vinfo, &def_stmt,
>         &def, &dt[1]))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "use not simple.");
>    return false;
> @@ -3586,7 +3586,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i
>        if (!vect_is_simple_use (op2, stmt, loop_vinfo, bb_vinfo, &def_stmt,
>         &def, &dt[2]))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "use not simple.");
>    return false;
> @@ -3628,7 +3628,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i
>        optab = optab_for_tree_code (code, vectype, optab_default);
>        if (!optab)
>   {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "no optab.");
>    return false;
> @@ -3638,14 +3638,14 @@ vectorizable_operation (gimple stmt, gimple_stmt_i
>
>    if (icode == CODE_FOR_nothing)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "op not supported by target.");
>        /* Check only during analysis.  */
>        if (GET_MODE_SIZE (vec_mode) != UNITS_PER_WORD
>    || (!vec_stmt && vf < vect_min_worthwhile_factor (code)))
>          return false;
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location, "proceeding using word mode.");
>      }
>
> @@ -3654,7 +3654,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i
>        && !vec_stmt
>        && vf < vect_min_worthwhile_factor (code))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "not worthwhile without SIMD support.");
>        return false;
> @@ -3663,7 +3663,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i
>    if (!vec_stmt) /* transformation not required.  */
>      {
>        STMT_VINFO_TYPE (stmt_info) = op_vec_info_type;
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>                           "=== vectorizable_operation ===");
>        vect_model_simple_cost (stmt_info, ncopies, dt, NULL, NULL);
> @@ -3672,7 +3672,7 @@ vectorizable_operation (gimple stmt, gimple_stmt_i
>
>    /** Transform.  **/
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "transform binary/unary operation.");
>
> @@ -3860,7 +3860,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera
>    /* FORNOW. This restriction should be relaxed.  */
>    if (loop && nested_in_vect_loop_p (loop, stmt) && ncopies > 1)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "multiple types in nested loop.");
>        return false;
> @@ -3894,7 +3894,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera
>    if (!vect_is_simple_use (op, stmt, loop_vinfo, bb_vinfo, &def_stmt,
>     &def, &dt))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "use not simple.");
>        return false;
> @@ -3915,7 +3915,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera
>      ? STMT_VINFO_DR_STEP (stmt_info) : DR_STEP (dr),
>      size_zero_node) < 0)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "negative step for store.");
>        return false;
> @@ -3946,7 +3946,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera
>                if (!vect_is_simple_use (op, next_stmt, loop_vinfo, bb_vinfo,
>         &def_stmt, &def, &dt))
>                  {
> -                  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                  if (dump_enabled_p ())
>                      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                       "use not simple.");
>                    return false;
> @@ -4008,7 +4008,7 @@ vectorizable_store (gimple stmt, gimple_stmt_itera
>        group_size = vec_num = 1;
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "transform store. ncopies = %d", ncopies);
>
> @@ -4396,7 +4396,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat
>    /* FORNOW. This restriction should be relaxed.  */
>    if (nested_in_vect_loop && ncopies > 1)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "multiple types in nested loop.");
>        return false;
> @@ -4436,7 +4436,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat
>      (e.g. - data copies).  */
>    if (optab_handler (mov_optab, mode) == CODE_FOR_nothing)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "Aligned load, but unsupported type.");
>        return false;
> @@ -4472,7 +4472,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat
>   &def_stmt, &def, &gather_dt,
>   &gather_off_vectype))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "gather index use not simple.");
>    return false;
> @@ -4492,7 +4492,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat
>         size_zero_node) < 0;
>        if (negative && ncopies > 1)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "multiple types with negative step.");
>    return false;
> @@ -4505,14 +4505,14 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat
>    if (alignment_support_scheme != dr_aligned
>        && alignment_support_scheme != dr_unaligned_supported)
>      {
> -              if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +              if (dump_enabled_p ())
>                  dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                   "negative step but alignment required.");
>        return false;
>      }
>    if (!perm_mask_for_reverse (vectype))
>      {
> -              if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +              if (dump_enabled_p ())
>                  dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                   "negative step and reversing not supported.");
>        return false;
> @@ -4527,7 +4527,7 @@ vectorizable_load (gimple stmt, gimple_stmt_iterat
>        return true;
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>                       "transform load. ncopies = %d", ncopies);
>
> @@ -5334,7 +5334,7 @@ vectorizable_condition (gimple stmt, gimple_stmt_i
>    /* FORNOW: not yet supported.  */
>    if (STMT_VINFO_LIVE_P (stmt_info))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "value used after loop.");
>        return false;
> @@ -5534,7 +5534,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect
>    gimple pattern_stmt;
>    gimple_seq pattern_def_seq;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "==> examining statement: ");
>        dump_gimple_stmt (MSG_NOTE, TDF_SLIM, stmt, 0);
> @@ -5542,7 +5542,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect
>
>    if (gimple_has_volatile_ops (stmt))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "not vectorized: stmt has volatile operands");
>
> @@ -5575,7 +5575,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect
>            /* Analyze PATTERN_STMT instead of the original stmt.  */
>            stmt = pattern_stmt;
>            stmt_info = vinfo_for_stmt (pattern_stmt);
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_NOTE, vect_location,
>                                 "==> examining pattern statement: ");
> @@ -5584,7 +5584,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect
>          }
>        else
>          {
> -          if (dump_kind_p (MSG_NOTE))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_NOTE, vect_location, "irrelevant.");
>
>            return true;
> @@ -5597,7 +5597,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect
>                 || STMT_VINFO_LIVE_P (vinfo_for_stmt (pattern_stmt))))
>      {
>        /* Analyze PATTERN_STMT too.  */
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "==> examining pattern statement: ");
> @@ -5621,7 +5621,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect
>        || STMT_VINFO_LIVE_P (vinfo_for_stmt (pattern_def_stmt)))
>      {
>        /* Analyze def stmt of STMT if it's a pattern stmt.  */
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_NOTE, vect_location,
>                                     "==> examining pattern def statement: ");
> @@ -5660,7 +5660,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect
>        gcc_assert (PURE_SLP_STMT (stmt_info));
>
>        scalar_type = TREE_TYPE (gimple_get_lhs (stmt));
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "get vectype for scalar type:  ");
> @@ -5670,7 +5670,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect
>        vectype = get_vectype_for_scalar_type (scalar_type);
>        if (!vectype)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                 dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                                  "not SLPed: unsupported data-type ");
> @@ -5680,7 +5680,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect
>            return false;
>          }
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_NOTE, vect_location, "vectype:  ");
>            dump_generic_expr (MSG_NOTE, TDF_SLIM, vectype);
> @@ -5724,7 +5724,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect
>
>    if (!ok)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                             "not vectorized: relevant stmt not ");
> @@ -5746,7 +5746,7 @@ vect_analyze_stmt (gimple stmt, bool *need_to_vect
>
>    if (!ok)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                             "not vectorized: live stmt not ");
> @@ -5846,7 +5846,7 @@ vect_transform_stmt (gimple stmt, gimple_stmt_iter
>      default:
>        if (!STMT_VINFO_LIVE_P (stmt_info))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                               "stmt not supported.");
>    gcc_unreachable ();
> @@ -5871,7 +5871,7 @@ vect_transform_stmt (gimple stmt, gimple_stmt_iter
>        tree scalar_dest;
>        gimple exit_phi;
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>                           "Record the vdef for outer-loop vectorization.");
>
> @@ -6108,7 +6108,7 @@ get_vectype_for_scalar_type_and_size (tree scalar_
>      return NULL_TREE;
>
>    vectype = build_vector_type (scalar_type, nunits);
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                         "get vectype with %d units of type ", nunits);
> @@ -6118,7 +6118,7 @@ get_vectype_for_scalar_type_and_size (tree scalar_
>    if (!vectype)
>      return NULL_TREE;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "vectype: ");
>        dump_generic_expr (MSG_NOTE, TDF_SLIM, vectype);
> @@ -6127,7 +6127,7 @@ get_vectype_for_scalar_type_and_size (tree scalar_
>    if (!VECTOR_MODE_P (TYPE_MODE (vectype))
>        && !INTEGRAL_MODE_P (TYPE_MODE (vectype)))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "mode not supported by target.");
>        return NULL_TREE;
> @@ -6198,7 +6198,7 @@ vect_is_simple_use (tree operand, gimple stmt, loo
>    *def_stmt = NULL;
>    *def = NULL_TREE;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location,
>                         "vect_is_simple_use: operand ");
> @@ -6220,14 +6220,14 @@ vect_is_simple_use (tree operand, gimple stmt, loo
>
>    if (TREE_CODE (operand) == PAREN_EXPR)
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location, "non-associatable copy.");
>        operand = TREE_OPERAND (operand, 0);
>      }
>
>    if (TREE_CODE (operand) != SSA_NAME)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "not ssa-name.");
>        return false;
> @@ -6236,13 +6236,13 @@ vect_is_simple_use (tree operand, gimple stmt, loo
>    *def_stmt = SSA_NAME_DEF_STMT (operand);
>    if (*def_stmt == NULL)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "no def_stmt.");
>        return false;
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "def_stmt: ");
>        dump_gimple_stmt (MSG_NOTE, TDF_SLIM, *def_stmt, 0);
> @@ -6274,13 +6274,13 @@ vect_is_simple_use (tree operand, gimple stmt, loo
>    && *dt == vect_double_reduction_def
>    && gimple_code (stmt) != GIMPLE_PHI))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "Unsupported pattern.");
>        return false;
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "type of def: %d.", *dt);
>
>    switch (gimple_code (*def_stmt))
> @@ -6299,7 +6299,7 @@ vect_is_simple_use (tree operand, gimple stmt, loo
>   break;
>        /* FALLTHRU */
>      default:
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "unsupported defining stmt: ");
>        return false;
> Index: tree-vect-slp.c
> ===================================================================
> --- tree-vect-slp.c (revision 192695)
> +++ tree-vect-slp.c (working copy)
> @@ -238,7 +238,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi
>         &def, &dt)
>    || (!def_stmt && dt != vect_constant_def))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>         "Build SLP failed: can't find def for ");
> @@ -263,7 +263,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi
>            pattern = true;
>            if (!first && !oprnd_info->first_pattern)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "Build SLP failed: some of the stmts"
> @@ -279,7 +279,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi
>
>            if (dt == vect_unknown_def_type)
>              {
> -              if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +              if (dump_enabled_p ())
>                  dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "Unsupported pattern.");
>                return false;
> @@ -296,7 +296,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi
>                  break;
>
>                default:
> -                if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                if (dump_enabled_p ())
>                    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "unsupported defining stmt: ");
>                  return false;
> @@ -361,7 +361,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi
>      {
>        if (number_of_oprnds != 2)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "Build SLP failed: different types ");
>
> @@ -388,7 +388,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi
>                             && !types_compatible_p (oprnd_info->first_def_type,
>                                                     TREE_TYPE (def_op0))))
>                      {
> -                      if (dump_kind_p (MSG_NOTE))
> +                      if (dump_enabled_p ())
>                  {
>    dump_printf_loc (MSG_NOTE, vect_location,
>     "Swapping operands of ");
> @@ -400,7 +400,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi
>      }
>                    else
>                      {
> -              if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +              if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "Build SLP failed: different types ");
>
> @@ -435,7 +435,7 @@ vect_get_and_check_slp_defs (loop_vec_info loop_vi
>
>   default:
>    /* FORNOW: Not supported.  */
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>         "Build SLP failed: illegal type of def ");
> @@ -504,7 +504,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>    /* For every stmt in NODE find its def stmt/s.  */
>    FOR_EACH_VEC_ELT (gimple, stmts, i, stmt)
>      {
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_NOTE, vect_location, "Build SLP for ");
>    dump_gimple_stmt (MSG_NOTE, TDF_SLIM, stmt, 0);
> @@ -513,7 +513,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>        /* Fail to vectorize statements marked as unvectorizable.  */
>        if (!STMT_VINFO_VECTORIZABLE (vinfo_for_stmt (stmt)))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>         "Build SLP failed: unvectorizable statement ");
> @@ -527,7 +527,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>        lhs = gimple_get_lhs (stmt);
>        if (lhs == NULL_TREE)
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>         "Build SLP failed: not GIMPLE_ASSIGN nor "
> @@ -544,7 +544,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>             && (cond = gimple_assign_rhs1 (stmt))
>             && !COMPARISON_CLASS_P (cond))
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>         "Build SLP failed: condition is not "
> @@ -560,7 +560,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>        vectype = get_vectype_for_scalar_type (scalar_type);
>        if (!vectype)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>         "Build SLP failed: unsupported data-type ");
> @@ -591,7 +591,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>        || !gimple_call_nothrow_p (stmt)
>        || gimple_call_chain (stmt))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "Build SLP failed: unsupported call type ");
> @@ -631,7 +631,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>
>    if (!optab)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "Build SLP failed: no optab.");
>          vect_free_oprnd_info (&oprnds_info);
> @@ -640,7 +640,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>    icode = (int) optab_handler (optab, vec_mode);
>    if (icode == CODE_FOR_nothing)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "Build SLP failed: "
>   "op not supported by target.");
> @@ -674,7 +674,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>                         || first_stmt_code == COMPONENT_REF
>                         || first_stmt_code == MEM_REF)))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "Build SLP failed: different operation "
> @@ -689,7 +689,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>    if (need_same_oprnds
>        && !operand_equal_p (first_op1, gimple_assign_rhs2 (stmt), 0))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "Build SLP failed: different shift "
> @@ -710,7 +710,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>    || gimple_call_fntype (first_stmt)
>       != gimple_call_fntype (stmt))
>   {
> -  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>         "Build SLP failed: different calls in ");
> @@ -749,7 +749,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>                    || (GROUP_FIRST_ELEMENT (vinfo_for_stmt (stmt)) != stmt
>                        && GROUP_GAP (vinfo_for_stmt (stmt)) != 1))
>                  {
> -                  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                  if (dump_enabled_p ())
>                      {
>                        dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>         "Build SLP failed: grouped "
> @@ -767,7 +767,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>                if (loop_vinfo
>                    && GROUP_SIZE (vinfo_for_stmt (stmt)) > ncopies * group_size)
>                  {
> -                  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                  if (dump_enabled_p ())
>                      {
>                        dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>         "Build SLP failed: the number "
> @@ -792,7 +792,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>                        && rhs_code != REALPART_EXPR
>                        && rhs_code != IMAGPART_EXPR)
>                      {
> -                      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                      if (dump_enabled_p ())
>                          {
>                            dump_printf_loc (MSG_MISSED_OPTIMIZATION,
>     vect_location,
> @@ -817,7 +817,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>                    if (vect_supportable_dr_alignment (first_dr, false)
>                        == dr_unaligned_unsupported)
>                      {
> -                      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                      if (dump_enabled_p ())
>                          {
>                            dump_printf_loc (MSG_MISSED_OPTIMIZATION,
>     vect_location,
> @@ -857,7 +857,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>    if (TREE_CODE_CLASS (rhs_code) == tcc_reference)
>      {
>        /* Not grouped load.  */
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "Build SLP failed: not grouped load ");
> @@ -875,7 +875,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>        && rhs_code != COND_EXPR
>        && rhs_code != CALL_EXPR)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "Build SLP failed: operation");
> @@ -895,7 +895,7 @@ vect_build_slp_tree (loop_vec_info loop_vinfo, bb_
>   first_cond_code = TREE_CODE (cond_expr);
>                else if (first_cond_code != TREE_CODE (cond_expr))
>                  {
> -                  if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +                  if (dump_enabled_p ())
>                      {
>                        dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>         "Build SLP failed: different"
> @@ -1080,7 +1080,7 @@ vect_supported_slp_permutation_p (slp_instance ins
>        /* Check that the loads are all in the same interleaving chain.  */
>        if (GROUP_FIRST_ELEMENT (vinfo_for_stmt (scalar_stmt)) != first_load)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>         "Build SLP failed: unsupported data "
> @@ -1169,7 +1169,7 @@ vect_supported_load_permutation_p (slp_instance sl
>    if (!slp_instn)
>      return false;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "Load permutation ");
>        FOR_EACH_VEC_ELT (int, load_permutation, i, next)
> @@ -1376,7 +1376,7 @@ vect_supported_load_permutation_p (slp_instance sl
>                    if (vect_supportable_dr_alignment (dr, false)
>                 == dr_unaligned_unsupported)
>                      {
> -        if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +        if (dump_enabled_p ())
>          {
>                      dump_printf_loc (MSG_MISSED_OPTIMIZATION,
>     vect_location,
> @@ -1536,7 +1536,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf
>
>    if (!vectype)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "Build SLP failed: unsupported data-type ");
> @@ -1556,7 +1556,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf
>    unrolling_factor = least_common_multiple (nunits, group_size) / group_size;
>    if (unrolling_factor != 1 && !loop_vinfo)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "Build SLP failed: unrolling required in basic"
>   " block SLP");
> @@ -1618,7 +1618,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf
>
>        if (unrolling_factor != 1 && !loop_vinfo)
>          {
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>       "Build SLP failed: unrolling required in basic"
>       " block SLP");
> @@ -1645,7 +1645,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf
>            if (!vect_supported_load_permutation_p (new_instance, group_size,
>                                                    load_permutation))
>              {
> -              if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +              if (dump_enabled_p ())
>                  {
>                    dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "Build SLP failed: unsupported load "
> @@ -1685,7 +1685,7 @@ vect_analyze_slp_instance (loop_vec_info loop_vinf
>          VEC_safe_push (slp_instance, heap, BB_VINFO_SLP_INSTANCES (bb_vinfo),
>                         new_instance);
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   vect_print_slp_tree (MSG_NOTE, node);
>
>        return true;
> @@ -1717,7 +1717,7 @@ vect_analyze_slp (loop_vec_info loop_vinfo, bb_vec
>    gimple first_element;
>    bool ok = false;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "=== vect_analyze_slp ===");
>
>    if (loop_vinfo)
> @@ -1736,7 +1736,7 @@ vect_analyze_slp (loop_vec_info loop_vinfo, bb_vec
>
>    if (bb_vinfo && !ok)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "Failed to SLP the basic block.");
>
> @@ -1780,7 +1780,7 @@ vect_make_slp_decision (loop_vec_info loop_vinfo)
>    slp_instance instance;
>    int decided_to_slp = 0;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "===
> vect_make_slp_decision ===");
>
>    FOR_EACH_VEC_ELT (slp_instance, slp_instances, i, instance)
> @@ -1798,7 +1798,7 @@ vect_make_slp_decision (loop_vec_info loop_vinfo)
>
>    LOOP_VINFO_SLP_UNROLLING_FACTOR (loop_vinfo) = unrolling_factor;
>
> -  if (decided_to_slp && dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (decided_to_slp && dump_enabled_p ())
>      dump_printf_loc (MSG_OPTIMIZED_LOCATIONS, vect_location,
>       "Decided to SLP %d instances. Unrolling factor %d",
>       decided_to_slp, unrolling_factor);
> @@ -1863,7 +1863,7 @@ vect_detect_hybrid_slp (loop_vec_info loop_vinfo)
>    VEC (slp_instance, heap) *slp_instances = LOOP_VINFO_SLP_INSTANCES
> (loop_vinfo);
>    slp_instance instance;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "===
> vect_detect_hybrid_slp ===");
>
>    FOR_EACH_VEC_ELT (slp_instance, slp_instances, i, instance)
> @@ -2060,7 +2060,7 @@ vect_bb_vectorization_profitable_p (bb_vec_info bb
>
>    vec_outside_cost = vec_prologue_cost + vec_epilogue_cost;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE, vect_location, "Cost model analysis: \n");
>        dump_printf (MSG_NOTE, "  Vector inside of basic block cost: %d\n",
> @@ -2097,7 +2097,7 @@ vect_slp_analyze_bb_1 (basic_block bb)
>
>    if (!vect_analyze_data_refs (NULL, bb_vinfo, &min_vf))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: unhandled data-ref in basic "
>   "block.\n");
> @@ -2109,7 +2109,7 @@ vect_slp_analyze_bb_1 (basic_block bb)
>    ddrs = BB_VINFO_DDRS (bb_vinfo);
>    if (!VEC_length (ddr_p, ddrs))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: not enough data-refs in "
>   "basic block.\n");
> @@ -2123,7 +2123,7 @@ vect_slp_analyze_bb_1 (basic_block bb)
>    if (!vect_analyze_data_ref_dependences (NULL, bb_vinfo, &max_vf)
>         || min_vf > max_vf)
>       {
> -       if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +       if (dump_enabled_p ())
>   dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>    "not vectorized: unhandled data dependence "
>    "in basic block.\n");
> @@ -2134,7 +2134,7 @@ vect_slp_analyze_bb_1 (basic_block bb)
>
>    if (!vect_analyze_data_refs_alignment (NULL, bb_vinfo))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: bad data alignment in basic "
>   "block.\n");
> @@ -2145,7 +2145,7 @@ vect_slp_analyze_bb_1 (basic_block bb)
>
>    if (!vect_analyze_data_ref_accesses (NULL, bb_vinfo))
>      {
> -     if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +     if (dump_enabled_p ())
>         dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: unhandled data access in "
>   "basic block.\n");
> @@ -2158,7 +2158,7 @@ vect_slp_analyze_bb_1 (basic_block bb)
>       trees.  */
>    if (!vect_analyze_slp (NULL, bb_vinfo))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: failed to find SLP opportunities "
>   "in basic block.\n");
> @@ -2179,7 +2179,7 @@ vect_slp_analyze_bb_1 (basic_block bb)
>
>    if (!vect_verify_datarefs_alignment (NULL, bb_vinfo))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>                           "not vectorized: unsupported alignment in basic "
>                           "block.\n");
> @@ -2189,7 +2189,7 @@ vect_slp_analyze_bb_1 (basic_block bb)
>
>    if (!vect_slp_analyze_operations (bb_vinfo))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: bad operation in basic block.\n");
>
> @@ -2201,7 +2201,7 @@ vect_slp_analyze_bb_1 (basic_block bb)
>    if (flag_vect_cost_model
>        && !vect_bb_vectorization_profitable_p (bb_vinfo))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: vectorization is not "
>   "profitable.\n");
> @@ -2210,7 +2210,7 @@ vect_slp_analyze_bb_1 (basic_block bb)
>        return NULL;
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>       "Basic block will be vectorized using SLP\n");
>
> @@ -2226,7 +2226,7 @@ vect_slp_analyze_bb (basic_block bb)
>    gimple_stmt_iterator gsi;
>    unsigned int vector_sizes;
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "===vect_slp_analyze_bb===\n");
>
>    for (gsi = gsi_start_bb (bb); !gsi_end_p (gsi); gsi_next (&gsi))
> @@ -2240,7 +2240,7 @@ vect_slp_analyze_bb (basic_block bb)
>
>    if (insns > PARAM_VALUE (PARAM_SLP_MAX_INSNS_IN_BB))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>   "not vectorized: too many instructions in "
>   "basic block.\n");
> @@ -2267,7 +2267,7 @@ vect_slp_analyze_bb (basic_block bb)
>
>        /* Try the next biggest vector size.  */
>        current_vector_size = 1 << floor_log2 (vector_sizes);
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          dump_printf_loc (MSG_NOTE, vect_location,
>   "***** Re-trying analysis with "
>   "vector size %d\n", current_vector_size);
> @@ -2292,7 +2292,7 @@ vect_update_slp_costs_according_to_vf (loop_vec_in
>    stmt_info_for_cost *si;
>    void *data = LOOP_VINFO_TARGET_COST_DATA (loop_vinfo);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location,
>       "=== vect_update_slp_costs_according_to_vf ===");
>
> @@ -2800,7 +2800,7 @@ vect_get_mask_element (gimple stmt, int first_mask
>       the next vector as well.  */
>    if (only_one_vec && *current_mask_element >= mask_nunits)
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "permutation requires at least two vectors ");
> @@ -2818,7 +2818,7 @@ vect_get_mask_element (gimple stmt, int first_mask
>            /* We either need the first vector too or have already moved to the
>               next vector. In both cases, this permutation needs three
>               vectors.  */
> -          if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +          if (dump_enabled_p ())
>              {
>                dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>         "permutation requires at "
> @@ -2884,7 +2884,7 @@ vect_transform_slp_perm_load (gimple stmt, VEC (tr
>
>    if (!can_vec_perm_p (mode, false, NULL))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_MISSED_OPTIMIZATION, vect_location,
>     "no vect permute for ");
> @@ -2964,7 +2964,7 @@ vect_transform_slp_perm_load (gimple stmt, VEC (tr
>
>    if (!can_vec_perm_p (mode, false, mask))
>      {
> -      if (dump_kind_p (MSG_MISSED_OPTIMIZATION))
> +      if (dump_enabled_p ())
>   {
>    dump_printf_loc (MSG_MISSED_OPTIMIZATION,
>     vect_location,
> @@ -3068,7 +3068,7 @@ vect_schedule_slp_instance (slp_tree node, slp_ins
>        SLP_TREE_NUMBER_OF_VEC_STMTS (node) = vec_stmts_size;
>      }
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      {
>        dump_printf_loc (MSG_NOTE,vect_location,
>         "------>vectorizing SLP node starting from: ");
> @@ -3177,7 +3177,7 @@ vect_schedule_slp (loop_vec_info loop_vinfo, bb_ve
>        /* Schedule the tree of INSTANCE.  */
>        is_store = vect_schedule_slp_instance (SLP_INSTANCE_TREE (instance),
>                                               instance, vf);
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>   dump_printf_loc (MSG_NOTE, vect_location,
>                           "vectorizing stmts using SLP.");
>      }
> @@ -3222,7 +3222,7 @@ vect_slp_transform_bb (basic_block bb)
>
>    gcc_assert (bb_vinfo);
>
> -  if (dump_kind_p (MSG_NOTE))
> +  if (dump_enabled_p ())
>      dump_printf_loc (MSG_NOTE, vect_location, "SLPing BB\n");
>
>    for (si = gsi_start_bb (bb); !gsi_end_p (si); gsi_next (&si))
> @@ -3230,7 +3230,7 @@ vect_slp_transform_bb (basic_block bb)
>        gimple stmt = gsi_stmt (si);
>        stmt_vec_info stmt_info;
>
> -      if (dump_kind_p (MSG_NOTE))
> +      if (dump_enabled_p ())
>          {
>            dump_printf_loc (MSG_NOTE, vect_location,
>                             "------>SLPing statement: ");
> @@ -3248,7 +3248,7 @@ vect_slp_transform_bb (basic_block bb)
>          }
>      }
>
> -  if (dump_kind_p (MSG_OPTIMIZED_LOCATIONS))
> +  if (dump_enabled_p ())
>      dump_printf (MSG_OPTIMIZED_LOCATIONS, "BASIC BLOCK VECTORIZED\n");
>
>    destroy_bb_vec_info (bb_vinfo);



More information about the Gcc-patches mailing list