[RFA] [V3] new pass for sign/zero extension elimination

Richard Sandiford richard.sandiford@arm.com
Wed Jan 3 12:07:54 GMT 2024


Jeff Law <jlaw@ventanamicro.com> writes:
> I know we're deep into stage3 and about to transition to stage4.  So if 
> the consensus is for this to wait, I'll understand....
>
> This it the V3 of the ext-dce patch based on Joern's work from last year.
>
> Changes since V2:
>    Handle MINUS
>    Minor logic cleanup for SUBREGs in ext_dce_process_sets
>    Includes Joern's carry_backpropagate work
>    Cleaned up and removed some use handling code for STRICT_LOW_PART
>    Moved non-local goto special case out of main use handling, similar to
>    how we handle CALL_INSN_FUSAGE
>    Use df_simple_dataflow rather than custom dataflow handling
>
> There's more cleanups we could be doing here, but the question is do we 
> stop commit what we've got and iterate on the trunk or do we defer until 
> gcc-15 in which case we iterate on a branch or something.

FWIW, I think it should go in now.  The patch looks good to me with the
changes below.

> This still is enabled at -O1 or above, but that's to get as much testing 
> as possible.  Assuming the rest is ACK'd for the trunk we'll put it into 
> the list of optimizations enabled by -O2.

Yeah, agree that -O2 seems like a better fit.

>
> 	PR target/95650
> 	PR rtl-optimization/96031
> 	PR rtl-optimization/104387
> 	PR rtl-optimization/111384
>
> gcc/
> 	* Makefile.in (OBJS): Add ext-dce.o.
> 	* common.opt (ext-dce): Add new option.
> 	* df-scan.cc (df_get_exit_block_use_set): No longer static.
> 	* df.h (df_get_exit_block_use_set): Prototype.
> 	* ext-dce.cc: New file.
> 	* passes.def: Add ext-dce before combine.
> 	* tree-pass.h (make_pass_ext_dce): Prototype..
>
> gcc/testsuite
> 	* gcc.target/riscv/core_bench_list.c: New test.
> 	* gcc.target/riscv/core_init_matrix.c: New test.
> 	* gcc.target/riscv/core_list_init.c: New test.
> 	* gcc.target/riscv/matrix_add_const.c: New test.
> 	* gcc.target/riscv/mem-extend.c: New test.
> 	* gcc.target/riscv/pr111384.c: New test.
>
>
> diff --git a/gcc/Makefile.in b/gcc/Makefile.in
> index 754eceb23bb..3450eb860c6 100644
> --- a/gcc/Makefile.in
> +++ b/gcc/Makefile.in
> @@ -1451,6 +1451,7 @@ OBJS = \
>  	explow.o \
>  	expmed.o \
>  	expr.o \
> +	ext-dce.o \
>  	fibonacci_heap.o \
>  	file-prefix-map.o \
>  	final.o \
> diff --git a/gcc/common.opt b/gcc/common.opt
> index d263a959df3..8bbcaad2ec4 100644
> --- a/gcc/common.opt
> +++ b/gcc/common.opt
> @@ -3846,4 +3846,8 @@ fipa-ra
>  Common Var(flag_ipa_ra) Optimization
>  Use caller save register across calls if possible.
>  
> +fext-dce
> +Common Var(flag_ext_dce, 1) Optimization Init(0)
> +Perform dead code elimination on zero and sign extensions with special dataflow analysis.
> +
>  ; This comment is to ensure we retain the blank line above.
> diff --git a/gcc/df-scan.cc b/gcc/df-scan.cc
> index 934c9ca2d81..93c0ba4e15c 100644
> --- a/gcc/df-scan.cc
> +++ b/gcc/df-scan.cc
> @@ -78,7 +78,6 @@ static void df_get_eh_block_artificial_uses (bitmap);
>  
>  static void df_record_entry_block_defs (bitmap);
>  static void df_record_exit_block_uses (bitmap);
> -static void df_get_exit_block_use_set (bitmap);
>  static void df_get_entry_block_def_set (bitmap);
>  static void df_grow_ref_info (struct df_ref_info *, unsigned int);
>  static void df_ref_chain_delete_du_chain (df_ref);
> @@ -3642,7 +3641,7 @@ df_epilogue_uses_p (unsigned int regno)
>  
>  /* Set the bit for regs that are considered being used at the exit. */
>  
> -static void
> +void
>  df_get_exit_block_use_set (bitmap exit_block_uses)
>  {
>    unsigned int i;
> diff --git a/gcc/df.h b/gcc/df.h
> index 402657a7076..abcbb097734 100644
> --- a/gcc/df.h
> +++ b/gcc/df.h
> @@ -1091,6 +1091,7 @@ extern bool df_epilogue_uses_p (unsigned int);
>  extern void df_set_regs_ever_live (unsigned int, bool);
>  extern void df_compute_regs_ever_live (bool);
>  extern void df_scan_verify (void);
> +extern void df_get_exit_block_use_set (bitmap);
>  
>  
>  /*----------------------------------------------------------------------------
> diff --git a/gcc/ext-dce.cc b/gcc/ext-dce.cc
> new file mode 100644
> index 00000000000..379264e0bca
> --- /dev/null
> +++ b/gcc/ext-dce.cc
> @@ -0,0 +1,964 @@
> +/* RTL dead zero/sign extension (code) elimination.
> +   Copyright (C) 2000-2022 Free Software Foundation, Inc.
> +
> +This file is part of GCC.
> +
> +GCC is free software; you can redistribute it and/or modify it under
> +the terms of the GNU General Public License as published by the Free
> +Software Foundation; either version 3, or (at your option) any later
> +version.
> +
> +GCC is distributed in the hope that it will be useful, but WITHOUT ANY
> +WARRANTY; without even the implied warranty of MERCHANTABILITY or
> +FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
> +for more details.
> +
> +You should have received a copy of the GNU General Public License
> +along with GCC; see the file COPYING3.  If not see
> +<http://www.gnu.org/licenses/>.  */
> +
> +#include "config.h"
> +#include "system.h"
> +#include "coretypes.h"
> +#include "backend.h"
> +#include "rtl.h"
> +#include "tree.h"
> +#include "memmodel.h"
> +#include "insn-config.h"
> +#include "emit-rtl.h"
> +#include "recog.h"
> +#include "cfganal.h"
> +#include "tree-pass.h"
> +#include "cfgrtl.h"
> +#include "rtl-iter.h"
> +#include "df.h"
> +#include "print-rtl.h"
> +
> +/* These should probably move into a C++ class.  */
> +static vec<bitmap_head> livein;
> +static bitmap all_blocks;
> +static bitmap livenow;
> +static bitmap changed_pseudos;
> +static bool modify;
> +
> +/* We consider four bit groups for liveness:
> +   bit 0..7   (least significant byte)
> +   bit 8..15  (second least significant byte)
> +   bit 16..31
> +   bit 32..BITS_PER_WORD-1  */
> +
> +/* Note this pass could be used to narrow memory loads too.  It's
> +   not clear if that's profitable or not in general.  */
> +
> +#define UNSPEC_P(X) (GET_CODE (X) == UNSPEC || GET_CODE (X) == UNSPEC_VOLATILE)
> +
> +/* If we know the destination of CODE only uses some low bits
> +   (say just the QI bits of an SI operation), then return true
> +   if we can propagate the need for just the subset of bits
> +   from the destination to the sources.
> +
> +   FIXME: This is safe for operands 1 and 2 of an IF_THEN_ELSE, but not
> +   operand 0.  Thus is likely would need some special casing to handle.  */
> +
> +static bool
> +safe_for_live_propagation (rtx_code code)
> +{
> +  /* First handle rtx classes which as a whole are known to
> +     be either safe or unsafe.  */
> +  switch (GET_RTX_CLASS (code))
> +    {
> +      case RTX_OBJ:
> +	return true;
> +
> +      case RTX_COMPARE:
> +      case RTX_COMM_COMPARE:
> +      case RTX_TERNARY:
> +	return false;
> +
> +      default:
> +	break;
> +    }
> +
> +  /* What's left are specific codes.  We only need to identify those
> +     which are safe.   */
> +  switch (code)
> +    {
> +    /* These are trivially safe.  */
> +    case SUBREG:
> +    case NOT:
> +    case ZERO_EXTEND:
> +    case SIGN_EXTEND:
> +    case TRUNCATE:
> +    case PLUS:
> +    case MINUS:
> +    case MULT:
> +    case SMUL_HIGHPART:
> +    case UMUL_HIGHPART:
> +    case AND:
> +    case IOR:
> +    case XOR:
> +      return true;
> +
> +    /* We can propagate for the shifted operand, but not the shift
> +       count.  The count is handled specially.  */
> +    case ASHIFT:
> +    case LSHIFTRT:
> +    case ASHIFTRT:
> +    case SS_ASHIFT:
> +    case US_ASHIFT:
> +      return true;
> +
> +    /* There may be other safe codes.  If so they can be added
> +       individually when discovered.  */
> +    default:
> +      return false;
> +    }
> +}
> +
> +/* Clear bits in LIVENOW and set bits in LIVE_TMP for objects
> +   set/clobbered by OBJ contained in INSN.
> +
> +   Conceptually it is always safe to ignore a particular destination
> +   here as that will result in more chunks of data being considered
> +   live.  That's what happens when we "continue" the main loop when
> +   we see something we don't know how to handle such as a vector
> +   mode destination.
> +
> +   The more accurate we are in identifying what objects (and chunks
> +   within an object) are set by INSN, the more aggressive the
> +   optimization phase during use handling will be.  */
> +
> +static void
> +ext_dce_process_sets (rtx_insn *insn, rtx obj, bitmap live_tmp)
> +{
> +  subrtx_iterator::array_type array;
> +  FOR_EACH_SUBRTX (iter, array, obj, NONCONST)
> +    {
> +      const_rtx x = *iter;
> +
> +      /* An EXPR_LIST (from call fusage) ends in NULL_RTX.  */
> +      if (x == NULL_RTX)
> +	continue;
> +
> +      if (UNSPEC_P (x))
> +	continue;
> +
> +      if (GET_CODE (x) == SET || GET_CODE (x) == CLOBBER)
> +	{
> +	  unsigned bit = 0;
> +	  x = SET_DEST (x);
> +
> +	  /* We don't support vector destinations or destinations
> +	     wider than DImode.  */
> +	  scalar_int_mode outer_mode;
> +	  if (!is_a <scalar_int_mode> (GET_MODE (x), &outer_mode)
> +	      || GET_MODE_BITSIZE (outer_mode) > 64)
> +	    {
> +	      /* Skip the subrtxs of this destination.  There is
> +		 little value in iterating into the subobjects, so
> +		 just skip them for a bit of efficiency.  */
> +	      iter.skip_subrtxes ();
> +	      continue;
> +	    }
> +
> +	  /* We could have (strict_low_part (subreg ...)).  We can not just
> +	     strip the STRICT_LOW_PART as that would result in clearing
> +	     some bits in LIVENOW that are still live.  So process the
> +	     STRICT_LOW_PART specially.  */
> +	  if (GET_CODE (x) == STRICT_LOW_PART)
> +	    {
> +	      x = XEXP (x, 0);
> +
> +	      /* The only valid operand of a STRICT_LOW_PART is a non
> +		 paradoxical SUBREG.  */
> +	      gcc_assert (SUBREG_P (x)
> +			  && !paradoxical_subreg_p (x)
> +			  && SUBREG_BYTE (x).is_constant ());
> +
> +	      /* I think we should always see a REG here.  But let's
> +		 be sure.  */
> +	      gcc_assert (REG_P (SUBREG_REG (x)));
> +
> +	      /* The inner mode might be larger, just punt for
> +		 that case.  Remember, we can not just continue to process
> +		 the inner RTXs due to the STRICT_LOW_PART.  */
> +	      if (!is_a <scalar_int_mode> (GET_MODE (SUBREG_REG (x)), &outer_mode)
> +		  || GET_MODE_BITSIZE (outer_mode) > 64)
> +		{
> +		  /* Skip the subrtxs of the STRICT_LOW_PART.  We can't
> +		     process them because it'll set objects as no longer
> +		     live when they are in fact still live.  */
> +		  iter.skip_subrtxes ();
> +		  continue;
> +		}
> +
> +	      /* Transfer all the LIVENOW bits for X into LIVE_TMP.  */
> +	      HOST_WIDE_INT rn = REGNO (SUBREG_REG (x));
> +	      for (HOST_WIDE_INT i = 4 * rn; i < 4 * rn + 4; i++)
> +		if (bitmap_bit_p (livenow, i))
> +		  bitmap_set_bit (live_tmp, i);
> +
> +	      /* The mode of the SUBREG tells us how many bits we can
> +		 clear.  */
> +	      machine_mode mode = GET_MODE (x);
> +	      HOST_WIDE_INT size
> +		= exact_log2 (GET_MODE_SIZE (mode).to_constant ()) + 1;
> +	      bitmap_clear_range (livenow, 4 * rn, size);
> +
> +	      /* We have fully processed this destination.  */
> +	      iter.skip_subrtxes ();
> +	      continue;
> +	    }
> +
> +	  /* Phase one of destination handling.  First remove any wrapper
> +	     such as SUBREG or ZERO_EXTRACT.  */
> +	  unsigned HOST_WIDE_INT mask
> +	    = GET_MODE_MASK (GET_MODE_INNER (GET_MODE (x)));
> +	  if (SUBREG_P (x))
> +	    {
> +	      /* If we have a SUBREG that is too wide, just continue the loop
> +		 and let the iterator go down into SUBREG_REG.  */
> +	      if (!is_a <scalar_int_mode> (GET_MODE (SUBREG_REG (x)), &outer_mode)
> +		  || GET_MODE_BITSIZE (outer_mode) > 64)
> +		continue;
> +
> +	      /* We can safely strip a paradoxical subreg.  The inner mode will
> +		 be narrower than the outer mode.  We'll clear fewer bits in
> +		 LIVENOW than we'd like, but that's always safe.  */
> +	      if (paradoxical_subreg_p (x))
> +		x = XEXP (x, 0);
> +	      else if (SUBREG_BYTE (x).is_constant ())
> +		{
> +		  bit = subreg_lsb (x).to_constant ();
> +		  mask = GET_MODE_MASK (GET_MODE (SUBREG_REG (x))) << bit;
> +		  gcc_assert (mask);
> +		  x = SUBREG_REG (x);
> +		}
> +	      else
> +		gcc_unreachable ();
> +	    }
> +
> +	  if (GET_CODE (x) == ZERO_EXTRACT)
> +	    {
> +	      /* If either the size or the start position is unknown,
> +		 then assume we know nothing about what is overwritten.
> +		 This is overly conservative, but safe.  */
> +	      if (!CONST_INT_P (XEXP (x, 1)) || !CONST_INT_P (XEXP (x, 2)))
> +		continue;
> +	      mask = (1ULL << INTVAL (XEXP (x, 1))) - 1;
> +	      bit = INTVAL (XEXP (x, 2));
> +	      if (BITS_BIG_ENDIAN)
> +		bit = (GET_MODE_BITSIZE (GET_MODE (x))
> +		       - INTVAL (XEXP (x, 1)) - bit).to_constant ();
> +	      x = XEXP (x, 0);

Should mask be shifted by bit here, like it is for the subregs?  E.g.:

  (set (zero_extract:SI R (const_int 2) (const_int 7)))

would currently give a mask of 0x3 and a bit of 7, whereas I think it
should be a mask of 0x180.  Without that we would only treat the low 8
bits of R as live, rather than the low 16 bits.

> +
> +	      /* We can certainly get (zero_extract (subreg ...)).  The
> +		 mode of the zero_extract and location should be sufficient
> +		 and we can just strip the SUBREG.  */
> +	      if (SUBREG_P (x))
> +		x = SUBREG_REG (x);
> +	    }
> +
> +	  /* BIT >= 64 indicates something went horribly wrong.  */
> +	  gcc_assert (bit <= 63);
> +
> +	  /* Now handle the actual object that was changed.  */
> +	  if (REG_P (x))
> +	    {
> +	      /* Transfer the appropriate bits from LIVENOW into
> +		 LIVE_TMP.  */
> +	      HOST_WIDE_INT rn = REGNO (x);
> +	      for (HOST_WIDE_INT i = 4 * rn; i < 4 * rn + 4; i++)
> +		if (bitmap_bit_p (livenow, i))
> +		  bitmap_set_bit (live_tmp, i);
> +
> +	      /* Now clear the bits known written by this instruction.
> +		 Note that BIT need not be a power of two, consider a
> +		 ZERO_EXTRACT destination.  */
> +	      int start = (bit < 8 ? 0 : bit < 16 ? 1 : bit < 32 ? 2 : 3);
> +	      int end = ((mask & ~0xffffffffULL) ? 4
> +			 : (mask & 0xffff0000ULL) ? 3
> +			 : (mask & 0xff00) ? 2 : 1);
> +	      bitmap_clear_range (livenow, 4 * rn + start, end - start);
> +	    }
> +	  /* Some ports generate (clobber (const_int)).  */
> +	  else if (CONST_INT_P (x))
> +	    continue;
> +	  else
> +	    gcc_assert (CALL_P (insn)
> +			|| MEM_P (x)
> +			|| x == pc_rtx
> +			|| GET_CODE (x) == SCRATCH);
> +
> +	  iter.skip_subrtxes ();
> +	}
> +      else if (GET_CODE (x) == COND_EXEC)
> +	{
> +	  /* This isn't ideal, but may not be so bad in practice.  */
> +	  iter.skip_subrtxes ();
> +	}
> +    }
> +}
> +
> +/* INSN has a sign/zero extended source inside SET that we will
> +   try to turn into a SUBREG.  */
> +static void
> +ext_dce_try_optimize_insn (rtx_insn *insn, rtx set)
> +{
> +  rtx src = SET_SRC (set);
> +  rtx inner = XEXP (src, 0);
> +
> +  /* Avoid (subreg (mem)) and other constructs which may be valid RTL, but
> +     not useful for this optimization.  */
> +  if (!REG_P (inner) && !SUBREG_P (inner))
> +    return;

It doesn't look like this does avoid (subreg (mem)) directly.  We'd need
to check REG_P (SUBREG_REG (inner)) for that.

> +
> +  rtx new_pattern;
> +  if (dump_file)
> +    {
> +      fprintf (dump_file, "Processing insn:\n");
> +      dump_insn_slim (dump_file, insn);
> +      fprintf (dump_file, "Trying to simplify pattern:\n");
> +      print_rtl_single (dump_file, SET_SRC (set));
> +    }
> +
> +  new_pattern = simplify_gen_subreg (GET_MODE (src), inner,
> +				     GET_MODE (inner), 0);
> +  /* simplify_gen_subreg may fail in which case NEW_PATTERN will be NULL.
> +     We must not pass that as a replacement pattern to validate_change.  */
> +  if (new_pattern)
> +    {
> +      int ok = validate_change (insn, &SET_SRC (set), new_pattern, false);
> +
> +      if (ok)
> +	bitmap_set_bit (changed_pseudos, REGNO (SET_DEST (set)));
> +
> +      if (dump_file)
> +	{
> +	  if (ok)
> +	    fprintf (dump_file, "Successfully transformed to:\n");
> +	  else
> +	    fprintf (dump_file, "Failed transformation to:\n");
> +
> +	  print_rtl_single (dump_file, new_pattern);
> +	  fprintf (dump_file, "\n");
> +	}
> +    }
> +  else
> +    {
> +      if (dump_file)
> +	fprintf (dump_file, "Unable to generate valid SUBREG expression.\n");
> +    }
> +}
> +
> +/* Some operators imply that their second operand is fully live,
> +   regardless of how many bits in the output are live.  An example
> +   would be the shift count on a target without SHIFT_COUNT_TRUNCATED
> +   defined.
> +
> +   Return TRUE if CODE is such an operator.  FALSE otherwise.  */
> +
> +static bool
> +binop_implies_op2_fully_live (rtx_code code)
> +{
> +  switch (code)
> +    {
> +    case ASHIFT:
> +    case LSHIFTRT:
> +    case ASHIFTRT:
> +    case ROTATE:
> +    case ROTATERT:
> +    case SS_ASHIFT:
> +    case US_ASHIFT:
> +      return !SHIFT_COUNT_TRUNCATED;
> +
> +    default:
> +      return false;
> +    }
> +}
> +
> +/* X, with code CODE, is an operation for which safe_for_live_propagation
> +   holds true, and bits set in MASK are live in the result.  Compute a
> +   mask of (potentially) live bits in the non-constant inputs.  In case of
> +   binop_implies_op2_fully_live (e.g. shifts), the computed mask may
> +   exclusively pertain to the first operand.  */
> +
> +HOST_WIDE_INT
> +carry_backpropagate (HOST_WIDE_INT mask, enum rtx_code code, rtx x)
> +{
> +  if (mask == 0)
> +    return 0;
> +
> +  enum machine_mode mode = GET_MODE_INNER (GET_MODE (x));
> +  HOST_WIDE_INT mmask = GET_MODE_MASK (mode);
> +  switch (code)
> +    {
> +    case PLUS:
> +    case MINUS:
> +    case MULT:
> +      return (2ULL << floor_log2 (mask)) - 1;
> +
> +    /* We propagate for the shifted operand, but not the shift
> +       count.  The count is handled specially.  */
> +    case ASHIFT:
> +      if (CONSTANT_P (XEXP (x, 1))
> +	  && known_lt (UINTVAL (XEXP (x, 1)), GET_MODE_BITSIZE (mode)))
> +	return mask >> INTVAL (XEXP (x, 1));
> +      return (2ULL << floor_log2 (mask)) - 1;
> +
> +    /* We propagate for the shifted operand, but not the shift
> +       count.  The count is handled specially.  */
> +    case LSHIFTRT:
> +      if (CONSTANT_P (XEXP (x, 1))
> +	  && known_lt (UINTVAL (XEXP (x, 1)), GET_MODE_BITSIZE (mode)))
> +	return mmask & (mask << INTVAL (XEXP (x, 1)));
> +      return mmask;
> +
> +    /* We propagate for the shifted operand, but not the shift
> +       count.  The count is handled specially.  */
> +    case ASHIFTRT:
> +      if (CONSTANT_P (XEXP (x, 1))
> +	  && known_lt (UINTVAL (XEXP (x, 1)), GET_MODE_BITSIZE (mode)))
> +	{
> +	  HOST_WIDE_INT sign = 0;
> +	  if (HOST_BITS_PER_WIDE_INT - clz_hwi (mask) + INTVAL (XEXP (x, 1))
> +	      > GET_MODE_BITSIZE (mode).to_constant ())
> +	    sign = 1ULL << (GET_MODE_BITSIZE (mode).to_constant () - 1);
> +	  return sign | (mmask & (mask << INTVAL (XEXP (x, 1))));
> +	}
> +      return mmask;
> +
> +    case SMUL_HIGHPART:
> +    case UMUL_HIGHPART:
> +      if (XEXP (x, 1) == const0_rtx)
> +	return 0;
> +      if (CONSTANT_P (XEXP (x, 1)))
> +	{
> +	  if (pow2p_hwi (INTVAL (XEXP (x, 1))))
> +	    return mmask & (mask << (GET_MODE_BITSIZE (mode).to_constant ()
> +				     - exact_log2 (INTVAL (XEXP (x, 1)))));

This will give a UB shift for x==1 (log2==0).  For that I guess the
mask should be the sign bit for SMUL_HIGHPART and 0 for UMUL_HIGHPART.
Or we could just punt to mmask, since we shouldn't see unfolded mul_highparts.

> +	  int bits = (HOST_BITS_PER_WIDE_INT
> +		      + GET_MODE_BITSIZE (mode).to_constant ()
> +		      - clz_hwi (mask) - ctz_hwi (INTVAL (XEXP (x, 1))));
> +	  if (bits < GET_MODE_BITSIZE (mode).to_constant ())
> +	    return (1ULL << bits) - 1;
> +	}
> +      return mmask;
> +
> +    case SIGN_EXTEND:
> +      if (mask & ~GET_MODE_MASK (GET_MODE_INNER (GET_MODE (XEXP (x, 0)))))
> +	mask |= 1ULL << (GET_MODE_BITSIZE (mode).to_constant () - 1);
> +      return mask;
> +
> +    /* We propagate for the shifted operand, but not the shift
> +       count.  The count is handled specially.  */
> +    case SS_ASHIFT:
> +    case US_ASHIFT:
> +      if (CONSTANT_P (XEXP (x, 1))
> +	  && UINTVAL (XEXP (x, 1)) < GET_MODE_BITSIZE (mode).to_constant ())
> +	{
> +	  return ((mmask & ~((unsigned HOST_WIDE_INT)mmask
> +			     >> (INTVAL (XEXP (x, 1))
> +				 + (XEXP (x, 1) != const0_rtx
> +				    && code == SS_ASHIFT))))
> +		  | (mask >> INTVAL (XEXP (x, 1))));
> +	}
> +      return mmask;
> +
> +    default:
> +      return mask;
> +    }
> +}
> +
> +/* Process uses in INSN contained in OBJ.  Set appropriate bits in LIVENOW
> +   for any chunks of pseudos that become live, potentially filtering using
> +   bits from LIVE_TMP.
> +
> +   If MODIFY is true, then optimize sign/zero extensions to SUBREGs when
> +   the extended bits are never read and mark pseudos which had extensions
> +   eliminated in CHANGED_PSEUDOS.  */
> +
> +static void
> +ext_dce_process_uses (rtx_insn *insn, rtx obj, bitmap live_tmp)
> +{
> +  subrtx_var_iterator::array_type array_var;
> +  FOR_EACH_SUBRTX_VAR (iter, array_var, obj, NONCONST)
> +    {
> +      /* An EXPR_LIST (from call fusage) ends in NULL_RTX.  */
> +      rtx x = *iter;
> +      if (x == NULL_RTX)
> +	continue;
> +
> +      /* So the basic idea in this FOR_EACH_SUBRTX_VAR loop is to
> +	 handle SETs explicitly, possibly propagating live information
> +	 into the uses.
> +
> +	 We may continue the loop at various points which will cause
> +	 iteration into the next level of RTL.  Breaking from the loop
> +	 is never safe as it can lead us to fail to process some of the
> +	 RTL and thus not make objects live when necessary.  */
> +      enum rtx_code xcode = GET_CODE (x);
> +      if (xcode == SET)
> +	{
> +	  const_rtx dst = SET_DEST (x);
> +	  rtx src = SET_SRC (x);
> +	  const_rtx y;
> +	  unsigned HOST_WIDE_INT bit = 0;
> +
> +	  /* The code of the RHS of a SET.  */
> +	  enum rtx_code code = GET_CODE (src);
> +
> +	  /* ?!? How much of this should mirror SET handling, potentially
> +	     being shared?   */
> +	  if (SUBREG_P (dst) && SUBREG_BYTE (dst).is_constant ())
> +	    {
> +	      bit = subreg_lsb (dst).to_constant ();
> +	      if (bit >= HOST_BITS_PER_WIDE_INT)
> +		bit = HOST_BITS_PER_WIDE_INT - 1;
> +	      dst = SUBREG_REG (dst);
> +	    }
> +	  else if (GET_CODE (dst) == ZERO_EXTRACT
> +		   || GET_CODE (dst) == STRICT_LOW_PART)
> +	    dst = XEXP (dst, 0);

Seems like we should handle the bit position in the ZERO_EXTRACT, since
if we look through to the first operand and still use live_tmp on it,
we're effectively assuming that XEXP (dst, 2) specifies the lsb.

But for now it might be simpler to remove ZERO_EXTRACT from the "else if"
and leave a FIXME, since continuing will be conservatively correct.

> +
> +	  /* Main processing of the uses.  Two major goals here.
> +
> +	     First, we want to try and propagate liveness (or the lack
> +	     thereof) from the destination register to the source
> +	     register(s).
> +
> +	     Second, if the source is an extension, try to optimize
> +	     it into a SUBREG.  The SUBREG form indicates we don't
> +	     care about the upper bits and will usually be copy
> +	     propagated away.
> +
> +	     If we fail to handle something in here, the expectation
> +	     is the iterator will dive into the sub-components and
> +	     mark all the chunks in any found REGs as live.  */
> +	  if (REG_P (dst) && safe_for_live_propagation (code))
> +	    {
> +	      /* Create a mask representing the bits of this output
> +		 operand that are live after this insn.  We can use
> +		 this information to refine the live in state of
> +		 inputs to this insn in many cases.
> +
> +		 We have to do this on a per SET basis, we might have
> +		 an INSN with multiple SETS, some of which can narrow
> +		 the source operand liveness, some of which may not.  */
> +	      unsigned HOST_WIDE_INT dst_mask = 0;
> +	      HOST_WIDE_INT rn = REGNO (dst);
> +	      unsigned HOST_WIDE_INT mask_array[]
> +		= { 0xff, 0xff00, 0xffff0000ULL, -0x100000000ULL };
> +	      for (int i = 0; i < 4; i++)
> +		if (bitmap_bit_p (live_tmp, 4 * rn + i))
> +		  dst_mask |= mask_array[i];
> +	      dst_mask >>= bit;
> +
> +	      /* ??? Could also handle ZERO_EXTRACT / SIGN_EXTRACT
> +		 of the source specially to improve optimization.  */
> +	      if (code == SIGN_EXTEND || code == ZERO_EXTEND)
> +		{
> +		  rtx inner = XEXP (src, 0);
> +		  unsigned HOST_WIDE_INT src_mask
> +		    = GET_MODE_MASK (GET_MODE_INNER (GET_MODE (inner)));
> +
> +		  /* DST_MASK could be zero if we had something in the SET
> +		     that we couldn't handle.  */
> +		  if (modify && dst_mask && (dst_mask & ~src_mask) == 0)
> +		    ext_dce_try_optimize_insn (insn, x);
> +
> +		  dst_mask &= src_mask;
> +		  src = XEXP (src, 0);
> +		  code = GET_CODE (src);
> +		}
> +
> +	      /* Optimization is done at this point.  We just want to make
> +		 sure everything that should get marked as live is marked
> +		 from here onward.  */
> +
> +	      dst_mask = carry_backpropagate (dst_mask, code, src);
> +
> +	      /* We will handle the other operand of a binary operator
> +		 at the bottom of the loop by resetting Y.  */
> +	      if (BINARY_P (src))
> +		y = XEXP (src, 0);
> +	      else
> +		y = src;
> +
> +	      /* We're inside a SET and want to process the source operands
> +		 making things live.  Breaking from this loop will cause
> +		 the iterator to work on sub-rtxs, so it is safe to break
> +		 if we see something we don't know how to handle.  */
> +	      for (;;)
> +		{
> +		  /* Strip an outer paradoxical subreg.  The bits outside
> +		     the inner mode are don't cares.  So we can just strip
> +		     and process the inner object.  */
> +		  if (paradoxical_subreg_p (y))
> +		    y = XEXP (y, 0);
> +		  else if (SUBREG_P (y) && SUBREG_BYTE (y).is_constant ())
> +		    {
> +		      /* For anything but (subreg (reg)), break the inner loop
> +			 and process normally (conservatively).  */
> +		      if (!REG_P (SUBREG_REG (y)))
> +			break;
> +		      bit = subreg_lsb (y).to_constant ();
> +		      if (dst_mask)
> +			{
> +			  dst_mask <<= bit;
> +			  if (!dst_mask)
> +			    dst_mask = -0x100000000ULL;
> +			}
> +		      y = SUBREG_REG (y);
> +		    }
> +
> +		  if (REG_P (y))
> +		    {
> +		      /* We have found the use of a register.  We need to mark
> +			 the appropriate chunks of the register live.  The mode
> +			 of the REG is a starting point.  We may refine that
> +			 based on what chunks in the output were live.  */
> +		      rn = 4 * REGNO (y);
> +		      unsigned HOST_WIDE_INT tmp_mask = dst_mask;
> +
> +		      /* If the RTX code for the SET_SRC is not one we can
> +			 propagate destination liveness through, then just
> +			 set the mask to the mode's mask.  */
> +		      if (!safe_for_live_propagation (code))
> +			tmp_mask
> +			  = GET_MODE_MASK (GET_MODE_INNER (GET_MODE (y)));
> +
> +		      if (tmp_mask & 0xff)
> +			bitmap_set_bit (livenow, rn);
> +		      if (tmp_mask & 0xff00)
> +			bitmap_set_bit (livenow, rn + 1);
> +		      if (tmp_mask & 0xffff0000ULL)
> +			bitmap_set_bit (livenow, rn + 2);
> +		      if (tmp_mask & -0x100000000ULL)
> +			bitmap_set_bit (livenow, rn + 3);
> +
> +		      /* Some operators imply their second operand
> +			 is fully live, break this inner loop which
> +			 will cause the iterator to descent into the
> +			 sub-rtxs outside the SET processing.  */
> +		      if (binop_implies_op2_fully_live (code))
> +			break;
> +		    }
> +		  else if (!CONSTANT_P (y))
> +		    break;
> +		  /* We might have (ashift (const_int 1) (reg...)) */
> +		  /* XXX share this logic with code below.  */
> +		  else if (binop_implies_op2_fully_live (GET_CODE (src)))
> +		    break;

I don't follow why we use GET_CODE (src) for constants and just "code"
for registers.  It shouldn't matter either way, since IIUC it's just a
question of whether this condition causes the break on the second iteration
(for operations where op2 *isn't* full live), or whether the later
!BINARY_P (src) does.  But I think it's significantly easier to follow
if we make the "else if" above a plain "if" and then delete the separate:

		      if (binop_implies_op2_fully_live (code))
			break;

[FTR, I know we both have misgivings about the iteration style, so I've
deliberately not commented on that this time.  But I think the above is
a nice localised change that improves readability, all IMO of course.]

> +
> +		  /* If this was anything but a binary operand, break the inner
> +		     loop.  This is conservatively correct as it will cause the
> +		     iterator to look at the sub-rtxs outside the SET context.  */
> +		  if (!BINARY_P (src))
> +		    break;
> +
> +		  /* We processed the first operand of a binary operator.  Now
> +		     handle the second.  */
> +		  y = XEXP (src, 1), src = pc_rtx;
> +		}
> +
> +	      /* These are leaf nodes, no need to iterate down into them.  */
> +	      if (REG_P (y) || CONSTANT_P (y))
> +		iter.skip_subrtxes ();
> +	    }
> +	}
> +      /* If we are reading the low part of a SUBREG, then we can
> +	 refine liveness of the input register, otherwise let the
> +	 iterator continue into SUBREG_REG.  */
> +      else if (SUBREG_P (x)
> +	       && REG_P (SUBREG_REG (x))
> +	       && subreg_lowpart_p (x)

Even paradoxical subregs are lowparts, so it might be worth adding

    !paradoxical_subreg_p (x)

here (for optimisation rather than correctness reasons).

> +	       && GET_MODE_BITSIZE (GET_MODE (x)).is_constant ()
> +	       && GET_MODE_BITSIZE (GET_MODE (x)).to_constant () <= 32)
> +	{
> +	  HOST_WIDE_INT size = GET_MODE_BITSIZE (GET_MODE  (x)).to_constant ();

Nit: excess space before "(x)"

> +	  HOST_WIDE_INT rn = 4 * REGNO (SUBREG_REG (x));
> +
> +	  bitmap_set_bit (livenow, rn);
> +	  if (size > 8)
> +	    bitmap_set_bit (livenow, rn + 1);
> +	  if (size > 16)
> +	    bitmap_set_bit (livenow, rn + 2);
> +	  if (size > 32)
> +	    bitmap_set_bit (livenow, rn + 3);

FWIW, this last part is dead code because of the "<= 32" above.
Might be worth turning into an assert instead.

> +	  iter.skip_subrtxes ();
> +	}
> +      /* If we have a register reference that is not otherwise handled,
> +	 just assume all the chunks are live.  */
> +      else if (REG_P (x))
> +	bitmap_set_range (livenow, REGNO (x) * 4, 4);
> +    }
> +}
> +
> +/* Process a single basic block BB with current liveness information
> +   in LIVENOW, returning updated liveness information.
> +
> +   If MODIFY is true, then this is the last pass and unnecessary
> +   extensions should be eliminated when possible.  If an extension
> +   is removed, the source pseudo is marked in CHANGED_PSEUDOS.  */
> +
> +static void
> +ext_dce_process_bb (basic_block bb)
> +{
> +  rtx_insn *insn;
> +
> +  FOR_BB_INSNS_REVERSE (bb, insn)
> +    {
> +      if (!NONDEBUG_INSN_P (insn))
> +	continue;
> +
> +      /* Live-out state of the destination of this insn.  We can
> +	 use this to refine the live-in state of the sources of
> +	 this insn in many cases.  */
> +      bitmap live_tmp = BITMAP_ALLOC (NULL);
> +
> +      /* First process any sets/clobbers in INSN.  */
> +      ext_dce_process_sets (insn, PATTERN (insn), live_tmp);
> +
> +      /* CALL_INSNs need processing their fusage data.  */
> +      if (CALL_P (insn))
> +	ext_dce_process_sets (insn, CALL_INSN_FUNCTION_USAGE (insn), live_tmp);
> +
> +      /* And now uses, optimizing away SIGN/ZERO extensions as we go.  */
> +      ext_dce_process_uses (insn, PATTERN (insn), live_tmp);
> +
> +      /* A nonlocal goto implicitly uses the frame pointer.  */
> +      if (JUMP_P (insn) && find_reg_note (insn, REG_NON_LOCAL_GOTO, NULL_RTX))
> +	{
> +	  bitmap_set_range (livenow, FRAME_POINTER_REGNUM * 4, 4);
> +	  if (!HARD_FRAME_POINTER_IS_FRAME_POINTER)
> +	    bitmap_set_range (livenow, HARD_FRAME_POINTER_REGNUM * 4, 4);
> +	}
> +
> +      /* And process fusage data for the use as well.  */
> +      if (CALL_P (insn))
> +	{
> +	  if (!FAKE_CALL_P (insn))
> +	    bitmap_set_range (livenow, STACK_POINTER_REGNUM * 4, 4);
> +
> +	  /* If this is not a call to a const fucntion, then assume it
> +	     can read any global register.  */
> +	  if (!RTL_CONST_CALL_P (insn))
> +	    for (unsigned i = 0; i < FIRST_PSEUDO_REGISTER; i++)
> +	      if (global_regs[i])
> +		bitmap_set_range (livenow, i * 4, 4);
> +
> +	  ext_dce_process_uses (insn, CALL_INSN_FUNCTION_USAGE (insn), live_tmp);
> +	}
> +
> +      BITMAP_FREE (live_tmp);
> +    }
> +}
> +
> +/* We optimize away sign/zero extensions in this pass and replace
> +   them with SUBREGs indicating certain bits are don't cares.
> +
> +   This changes the SUBREG_PROMOTED_VAR_P state of the object.
> +   It is fairly painful to fix this on the fly, so we have
> +   recorded which pseudos are affected and we look for SUBREGs
> +   of those pseudos and fix them up.  */
> +
> +static void
> +reset_subreg_promoted_p (void)
> +{
> +  /* If we removed an extension, that changed the promoted state
> +     of the destination of that extension.  Thus we need to go
> +     find any SUBREGs that reference that pseudo and adjust their
> +     SUBREG_PROMOTED_P state.  */
> +  for (rtx_insn *insn = get_insns(); insn; insn = NEXT_INSN (insn))
> +    {
> +      if (!NONDEBUG_INSN_P (insn))
> +	continue;
> +
> +      rtx pat = PATTERN (insn);
> +      subrtx_var_iterator::array_type array;
> +      FOR_EACH_SUBRTX_VAR (iter, array, pat, NONCONST)
> +	{
> +	  rtx sub = *iter;
> +
> +	  /* We only care about SUBREGs.  */
> +	  if (GET_CODE (sub) != SUBREG)
> +	    continue;
> +
> +	  const_rtx x = SUBREG_REG (sub);
> +
> +	  /* We only care if the inner object is a REG.  */
> +	  if (!REG_P (x))
> +	    continue;
> +
> +	  /* And only if the SUBREG is a promoted var.  */
> +	  if (!SUBREG_PROMOTED_VAR_P (sub))
> +	    continue;
> +
> +	  if (bitmap_bit_p (changed_pseudos, REGNO (x)))
> +	    SUBREG_PROMOTED_VAR_P (sub) = 0;
> +	}
> +    }
> +}
> +
> +/* Initialization of the ext-dce pass.  Primarily this means
> +   setting up the various bitmaps we utilize.  */
> +
> +static void
> +ext_dce_init (void)
> +{
> +

Nit: excess blank line.

> +  livein.create (last_basic_block_for_fn (cfun));
> +  livein.quick_grow_cleared (last_basic_block_for_fn (cfun));
> +  for (int i = 0; i < last_basic_block_for_fn (cfun); i++)
> +    bitmap_initialize (&livein[i], &bitmap_default_obstack);
> +
> +  auto_bitmap refs (&bitmap_default_obstack);
> +  df_get_exit_block_use_set (refs);
> +
> +  unsigned i;
> +  bitmap_iterator bi;
> +  EXECUTE_IF_SET_IN_BITMAP (refs, 0, i, bi)
> +    {
> +      for (int j = 0; j < 4; j++)
> +	bitmap_set_bit (&livein[EXIT_BLOCK], i * 4 + j);
> +    }
> +
> +  livenow = BITMAP_ALLOC (NULL);
> +  all_blocks = BITMAP_ALLOC (NULL);
> +  changed_pseudos = BITMAP_ALLOC (NULL);
> +
> +  for (int i = 0; i < n_basic_blocks_for_fn (cfun); i++)
> +    if (i != ENTRY_BLOCK && i != EXIT_BLOCK)
> +      bitmap_set_bit (all_blocks, i);
> +
> +  modify = false;
> +

Same here

> +}
> +
> +/* Finalization of the ext-dce pass.  Primarily this means
> +   releasing up the various bitmaps we utilize.  */
> +
> +static void
> +ext_dce_finish (void)
> +{
> +

Here too.

> +  for (unsigned i = 0; i < livein.length (); i++)
> +    bitmap_clear (&livein[i]);
> +  livein.release ();
> +
> +  BITMAP_FREE (livenow);
> +  BITMAP_FREE (changed_pseudos);
> +  BITMAP_FREE (all_blocks);
> +}
> +
> +/* Process block number BB_INDEX as part of the backward
> +   simple dataflow analysis.  Return TRUE if something in
> +   this block change or FALSE otherwise.  */

s/change/changed/

> +
> +static bool
> +ext_dce_rd_transfer_n (int bb_index)
> +{
> +  /* The ENTRY/EXIT blocks never change.  */
> +  if (bb_index == ENTRY_BLOCK || bb_index == EXIT_BLOCK)
> +    return false;
> +
> +  basic_block bb = BASIC_BLOCK_FOR_FN (cfun, bb_index);
> +
> +  /* Make everything live that's live in the successors.  */
> +  bitmap_clear (livenow);
> +  edge_iterator ei;
> +  edge e;
> +
> +  FOR_EACH_EDGE (e, ei, bb->succs)
> +    bitmap_ior_into (livenow, &livein[e->dest->index]);
> +
> +  ext_dce_process_bb (bb);
> +
> +  /* We may have narrowed the set of live objects at the start
> +     of this block.  If so, update the bitmaps and indicate to
> +     the generic dataflow code that something changed.  */
> +  if (!bitmap_equal_p (&livein[bb_index], livenow))
> +    {
> +      bitmap tmp = BITMAP_ALLOC (NULL);
> +      gcc_assert (!bitmap_and_compl (tmp, &livein[bb_index], livenow));
> +      BITMAP_FREE (tmp);

There's a bitmap_intersect_compl_p that would avoid the temporary.

> +
> +      bitmap_copy (&livein[bb_index], livenow);
> +      return true;
> +    }
> +
> +  return false;
> +}
> +
> +/* Dummy function for the df_simple_dataflow API.  */
> +static bool ext_dce_rd_confluence_n (edge) { return true; }
> +
> +/* Use lifetime analyis to identify extensions that set bits that
> +   are never read.  Turn such extensions into SUBREGs instead which
> +   can often be propagated away.  */
> +
> +static void
> +ext_dce (void)
> +{
> +
> +  df_analyze ();
> +  ext_dce_init ();
> +
> +  do
> +    {
> +      df_simple_dataflow (DF_BACKWARD, NULL, NULL,
> +			  ext_dce_rd_confluence_n, ext_dce_rd_transfer_n,
> +			  all_blocks, df_get_postorder (DF_BACKWARD),
> +			  df_get_n_blocks (DF_BACKWARD));
> +      modify = !modify;
> +    }
> +  while (modify);
> +
> +  reset_subreg_promoted_p ();
> +
> +  ext_dce_finish ();
> +

Nit: excess blank lines around function body.

Thanks,
Richard

> +}
> +
> +
> +namespace {
> +
> +const pass_data pass_data_ext_dce =
> +{
> +  RTL_PASS, /* type */
> +  "ext_dce", /* name */
> +  OPTGROUP_NONE, /* optinfo_flags */
> +  TV_NONE, /* tv_id */
> +  PROP_cfglayout, /* properties_required */
> +  0, /* properties_provided */
> +  0, /* properties_destroyed */
> +  0, /* todo_flags_start */
> +  TODO_df_finish, /* todo_flags_finish */
> +};
> +
> +class pass_ext_dce : public rtl_opt_pass
> +{
> +public:
> +  pass_ext_dce (gcc::context *ctxt)
> +    : rtl_opt_pass (pass_data_ext_dce, ctxt)
> +  {}
> +
> +  /* opt_pass methods: */
> +  virtual bool gate (function *) { return optimize > 0; }
> +  virtual unsigned int execute (function *)
> +    {
> +      ext_dce ();
> +      return 0;
> +    }
> +
> +}; // class pass_combine
> +
> +} // anon namespace
> +
> +rtl_opt_pass *
> +make_pass_ext_dce (gcc::context *ctxt)
> +{
> +  return new pass_ext_dce (ctxt);
> +}
> diff --git a/gcc/passes.def b/gcc/passes.def
> index 43b416f98f7..48bde575cf0 100644
> --- a/gcc/passes.def
> +++ b/gcc/passes.def
> @@ -491,6 +491,7 @@ along with GCC; see the file COPYING3.  If not see
>        NEXT_PASS (pass_inc_dec);
>        NEXT_PASS (pass_initialize_regs);
>        NEXT_PASS (pass_ud_rtl_dce);
> +      NEXT_PASS (pass_ext_dce);
>        NEXT_PASS (pass_combine);
>        NEXT_PASS (pass_if_after_combine);
>        NEXT_PASS (pass_jump_after_combine);
> diff --git a/gcc/testsuite/gcc.target/riscv/core_bench_list.c b/gcc/testsuite/gcc.target/riscv/core_bench_list.c
> new file mode 100644
> index 00000000000..957e9c841ed
> --- /dev/null
> +++ b/gcc/testsuite/gcc.target/riscv/core_bench_list.c
> @@ -0,0 +1,15 @@
> +/* { dg-do compile } */
> +/* { dg-options "-O2 -fdump-rtl-ext_dce" } */
> +/* { dg-final { scan-rtl-dump {Successfully transformed} "ext_dce" } } */
> +
> +short
> +core_bench_list (int N) {
> +
> +  short a = 0;
> +  for (int i = 0; i < 4; i++) {
> +    if (i > N) {
> +      a++;
> +    }
> +  }
> +  return a * 4;
> +}
> diff --git a/gcc/testsuite/gcc.target/riscv/core_init_matrix.c b/gcc/testsuite/gcc.target/riscv/core_init_matrix.c
> new file mode 100644
> index 00000000000..9289244c71f
> --- /dev/null
> +++ b/gcc/testsuite/gcc.target/riscv/core_init_matrix.c
> @@ -0,0 +1,17 @@
> +/* { dg-do compile } */
> +/* { dg-options "-O1 -fdump-rtl-ext_dce" } */
> +/* { dg-final { scan-rtl-dump {Successfully transformed} "ext_dce" } } */
> +
> +void
> +core_init_matrix(short* A, short* B, int seed) {
> +  int  order = 1;
> +
> +  for (int i = 0; i < seed; i++) {
> +    for (int j = 0; j < seed; j++) {
> +      short val = seed + order;
> +      B[i] = val;
> +      A[i] = val;
> +      order++;
> +    }
> +  }
> +}
> diff --git a/gcc/testsuite/gcc.target/riscv/core_list_init.c b/gcc/testsuite/gcc.target/riscv/core_list_init.c
> new file mode 100644
> index 00000000000..2f36dae85aa
> --- /dev/null
> +++ b/gcc/testsuite/gcc.target/riscv/core_list_init.c
> @@ -0,0 +1,18 @@
> +/* { dg-do compile } */
> +/* { dg-options "-O1 -fdump-rtl-ext_dce" } */
> +/* { dg-final { scan-rtl-dump {Successfully transformed} "ext_dce" } } */
> +
> +unsigned short
> +core_list_init (int size, short seed) {
> +
> +  for (int i = 0; i < size; i++) {
> +    unsigned short datpat = ((unsigned short)(seed ^ i) & 0xf);
> +    unsigned short dat = (datpat << 3) | (i & 0x7);
> +    if (i > seed) {
> +      return dat;
> +    }
> +  }
> +
> +  return 0;
> +
> +}
> diff --git a/gcc/testsuite/gcc.target/riscv/matrix_add_const.c b/gcc/testsuite/gcc.target/riscv/matrix_add_const.c
> new file mode 100644
> index 00000000000..9a2dd53b17a
> --- /dev/null
> +++ b/gcc/testsuite/gcc.target/riscv/matrix_add_const.c
> @@ -0,0 +1,11 @@
> +/* { dg-do compile } */
> +/* { dg-options "-O2 -fdump-rtl-ext_dce" } */
> +/* { dg-final { scan-rtl-dump {Successfully transformed} "ext_dce" } } */
> +
> +void
> +matrix_add_const(int N, short *A, short val)
> +{
> +    for (int j = 0; j < N; j++) {
> +      A[j] += val;
> +    }
> +}
> diff --git a/gcc/testsuite/gcc.target/riscv/mem-extend.c b/gcc/testsuite/gcc.target/riscv/mem-extend.c
> new file mode 100644
> index 00000000000..c67f12dfc35
> --- /dev/null
> +++ b/gcc/testsuite/gcc.target/riscv/mem-extend.c
> @@ -0,0 +1,13 @@
> +/* { dg-do compile } */
> +/* { dg-options "-march=rv64gc_zbb" } */
> +/* { dg-skip-if "" { *-*-* } { "-O0" } } */
> +
> +void
> +foo(short *d, short *tmp) {
> +    int x = d[0] + d[1];
> +    int y = d[2] + d[3];
> +    tmp[0] = x + y;
> +    tmp[1] = x - y;
> +}
> +
> +/* { dg-final { scan-assembler-not {\mzext\.h\M} } } */
> diff --git a/gcc/testsuite/gcc.target/riscv/pr111384.c b/gcc/testsuite/gcc.target/riscv/pr111384.c
> new file mode 100644
> index 00000000000..a4e77d4aeb6
> --- /dev/null
> +++ b/gcc/testsuite/gcc.target/riscv/pr111384.c
> @@ -0,0 +1,11 @@
> +/* { dg-do compile } */
> +/* { dg-options "-O1 -fdump-rtl-ext_dce" } */
> +/* { dg-final { scan-rtl-dump {Successfully transformed} "ext_dce" } } */
> +
> +void
> +foo(unsigned int src, unsigned short *dst1, unsigned short *dst2)
> +{
> +    *dst1 = src;
> +    *dst2 = src;
> +}
> +
> diff --git a/gcc/tree-pass.h b/gcc/tree-pass.h
> index 52fd57fd4c6..501501560ec 100644
> --- a/gcc/tree-pass.h
> +++ b/gcc/tree-pass.h
> @@ -593,6 +593,7 @@ extern rtl_opt_pass *make_pass_reginfo_init (gcc::context *ctxt);
>  extern rtl_opt_pass *make_pass_inc_dec (gcc::context *ctxt);
>  extern rtl_opt_pass *make_pass_stack_ptr_mod (gcc::context *ctxt);
>  extern rtl_opt_pass *make_pass_initialize_regs (gcc::context *ctxt);
> +extern rtl_opt_pass *make_pass_ext_dce (gcc::context *ctxt);
>  extern rtl_opt_pass *make_pass_combine (gcc::context *ctxt);
>  extern rtl_opt_pass *make_pass_if_after_combine (gcc::context *ctxt);
>  extern rtl_opt_pass *make_pass_jump_after_combine (gcc::context *ctxt);


More information about the Gcc-patches mailing list