https://gcc.gnu.org/bugzilla/show_bug.cgi?id=114009
Jakub Jelinek changed:
What|Removed |Added
Status|ASSIGNED|RESOLVED
Resolution|---
https://gcc.gnu.org/bugzilla/show_bug.cgi?id=114009
--- Comment #5 from GCC Commits ---
The master branch has been updated by Jakub Jelinek :
https://gcc.gnu.org/g:95b6ee96348041eaee9133f082b57f3e57ef0b11
commit r14-9350-g95b6ee96348041eaee9133f082b57f3e57ef0b11
Author: Jakub Jelinek
Date:
https://gcc.gnu.org/bugzilla/show_bug.cgi?id=114009
Jakub Jelinek changed:
What|Removed |Added
Status|NEW |ASSIGNED
Assignee|unassigned
https://gcc.gnu.org/bugzilla/show_bug.cgi?id=114009
--- Comment #3 from Jakub Jelinek ---
That said, I fail to see why the a/2*2 in there matters.
a*!a is simply always 0 for integral types, both signed and unsigned, including
signed 1-bit precision. If a is 0, the result is 0*1 (or for the
https://gcc.gnu.org/bugzilla/show_bug.cgi?id=114009
Jakub Jelinek changed:
What|Removed |Added
CC||jakub at gcc dot gnu.org
--- Comment
https://gcc.gnu.org/bugzilla/show_bug.cgi?id=114009
Richard Biener changed:
What|Removed |Added
Blocks||85316
Priority|P3
https://gcc.gnu.org/bugzilla/show_bug.cgi?id=114009
Andrew Pinski changed:
What|Removed |Added
Status|UNCONFIRMED |NEW
Last reconfirmed|
https://gcc.gnu.org/bugzilla/show_bug.cgi?id=114009
Andrew Pinski changed:
What|Removed |Added
Target Milestone|--- |11.5
Known to work|