This is the mail archive of the
gcc-bugs@gcc.gnu.org
mailing list for the GCC project.
[Bug tree-optimization/24599] [4.0 regression] segv after overflow detection
- From: "pinskia at gcc dot gnu dot org" <gcc-bugzilla at gcc dot gnu dot org>
- To: gcc-bugs at gcc dot gnu dot org
- Date: 3 Nov 2005 18:25:19 -0000
- Subject: [Bug tree-optimization/24599] [4.0 regression] segv after overflow detection
- References: <bug-24599-7786@http.gcc.gnu.org/bugzilla/>
- Reply-to: gcc-bugzilla at gcc dot gnu dot org
------- Comment #3 from pinskia at gcc dot gnu dot org 2005-11-03 18:25 -------
(In reply to comment #2)
> I'm not sure the root cause of this bug is fixed in 4.1. It looks to me like
> it's still there and is only (currently) hidden. Am I mistaken?
> Apply this patch to gcc-4.1-20051029 and recompile the testcase with -O3.
> I'm seeing an abort. If I set a breakpoint on integer_nonzerop and print the
> boolean_true_node tree I see it's marked as "overflow". Oops
Overflow should mean nothing to the optimizers or expanders. It is only a
language term really that should only matter to the front-end.
integer_nonzerop should not be used that much any more, Yes I know it is but
really it needs to be changed to use nonzero_p instead.
--
http://gcc.gnu.org/bugzilla/show_bug.cgi?id=24599