From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.129.124]) by sourceware.org (Postfix) with ESMTPS id 04D813858D37 for ; Thu, 3 Feb 2022 07:56:44 +0000 (GMT) DMARC-Filter: OpenDMARC Filter v1.4.1 sourceware.org 04D813858D37 Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-41-LD9IVJ1TNw6FgOuk8dJxdQ-1; Thu, 03 Feb 2022 02:56:42 -0500 X-MC-Unique: LD9IVJ1TNw6FgOuk8dJxdQ-1 Received: from smtp.corp.redhat.com (int-mx01.intmail.prod.int.phx2.redhat.com [10.5.11.11]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id 3A6E71853028; Thu, 3 Feb 2022 07:56:41 +0000 (UTC) Received: from tucnak.zalov.cz (unknown [10.39.192.125]) by smtp.corp.redhat.com (Postfix) with ESMTPS id 1D2AB5DBBC; Thu, 3 Feb 2022 07:56:37 +0000 (UTC) Received: from tucnak.zalov.cz (localhost [127.0.0.1]) by tucnak.zalov.cz (8.16.1/8.16.1) with ESMTPS id 2137uZki3856271 (version=TLSv1.3 cipher=TLS_AES_256_GCM_SHA384 bits=256 verify=NOT); Thu, 3 Feb 2022 08:56:35 +0100 Received: (from jakub@localhost) by tucnak.zalov.cz (8.16.1/8.16.1/Submit) id 2137uY1F3856270; Thu, 3 Feb 2022 08:56:34 +0100 Date: Thu, 3 Feb 2022 08:56:34 +0100 From: Jakub Jelinek To: Richard Biener , Andrew MacLeod Cc: gcc-patches@gcc.gnu.org Subject: [PATCH] ranger: Fix up wi_fold_in_parts for small precision types [PR104334] Message-ID: <20220203075634.GF2646553@tucnak> Reply-To: Jakub Jelinek MIME-Version: 1.0 X-Scanned-By: MIMEDefang 2.79 on 10.5.11.11 X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Type: text/plain; charset=us-ascii Content-Disposition: inline X-Spam-Status: No, score=-0.5 required=5.0 tests=BAYES_00, DKIMWL_WL_HIGH, DKIM_SIGNED, DKIM_VALID, DKIM_VALID_AU, DKIM_VALID_EF, KAM_SOMETLD_ARE_BAD_TLD, RCVD_IN_DNSWL_LOW, SPF_HELO_NONE, SPF_NONE, TXREP, T_SCC_BODY_TEXT_LINE autolearn=no autolearn_force=no version=3.4.4 X-Spam-Checker-Version: SpamAssassin 3.4.4 (2020-01-24) on server2.sourceware.org X-BeenThere: gcc-patches@gcc.gnu.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: Gcc-patches mailing list List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Thu, 03 Feb 2022 07:56:46 -0000 Hi! The wide-int.h templates expect that when an int/long etc. operand is used it will be sign-extended based on the types precision. wi_fold_in_parts passes 3 such non-zero constants to wi::lt_p, wi::gt_p and wi::eq_p - 1, 3 and 4, which means it was doing weird things if either some of 1, 3 or 4 weren't representable in type, or if type was unsigned 3 bit type 4 should be written as -4. The following patch promotes the subtraction operands to widest_int and uses that as the type for ?h_range variables and compares them as such. We don't need the overflow handling because there is never an overflow. Bootstrapped/regtested on x86_64-linux and i686-linux (on the former also with lto bootstrap), ok for trunk? 2022-02-02 Jakub Jelinek PR tree-optimization/104334 * range-op.cc (range_operator::wi_fold_in_parts): Change lh_range and rh_range type to widest_int and subtract in widest_int. Remove ov_rh, ov_lh and sign vars, always perform comparisons as signed and use >, < and == operators for it. * g++.dg/opt/pr104334.C: New test. --- gcc/range-op.cc.jj 2022-01-13 22:29:15.345831749 +0100 +++ gcc/range-op.cc 2022-02-02 20:20:22.020000000 +0100 @@ -144,22 +144,21 @@ range_operator::wi_fold_in_parts (irange const wide_int &rh_lb, const wide_int &rh_ub) const { - wi::overflow_type ov_rh, ov_lh; int_range_max tmp; - wide_int rh_range = wi::sub (rh_ub, rh_lb, TYPE_SIGN (type), &ov_rh); - wide_int lh_range = wi::sub (lh_ub, lh_lb, TYPE_SIGN (type), &ov_lh); - signop sign = TYPE_SIGN (type);; + widest_int rh_range = wi::sub (widest_int::from (rh_ub, TYPE_SIGN (type)), + widest_int::from (rh_lb, TYPE_SIGN (type))); + widest_int lh_range = wi::sub (widest_int::from (lh_ub, TYPE_SIGN (type)), + widest_int::from (lh_lb, TYPE_SIGN (type))); // If there are 2, 3, or 4 values in the RH range, do them separately. // Call wi_fold_in_parts to check the RH side. - if (wi::gt_p (rh_range, 0, sign) && wi::lt_p (rh_range, 4, sign) - && ov_rh == wi::OVF_NONE) + if (rh_range > 0 && rh_range < 4) { wi_fold_in_parts (r, type, lh_lb, lh_ub, rh_lb, rh_lb); - if (wi::gt_p (rh_range, 1, sign)) + if (rh_range > 1) { wi_fold_in_parts (tmp, type, lh_lb, lh_ub, rh_lb + 1, rh_lb + 1); r.union_ (tmp); - if (wi::eq_p (rh_range, 3)) + if (rh_range == 3) { wi_fold_in_parts (tmp, type, lh_lb, lh_ub, rh_lb + 2, rh_lb + 2); r.union_ (tmp); @@ -170,15 +169,14 @@ range_operator::wi_fold_in_parts (irange } // Otherise check for 2, 3, or 4 values in the LH range and split them up. // The RH side has been checked, so no recursion needed. - else if (wi::gt_p (lh_range, 0, sign) && wi::lt_p (lh_range, 4, sign) - && ov_lh == wi::OVF_NONE) + else if (lh_range > 0 && lh_range < 4) { wi_fold (r, type, lh_lb, lh_lb, rh_lb, rh_ub); - if (wi::gt_p (lh_range, 1, sign)) + if (lh_range > 1) { wi_fold (tmp, type, lh_lb + 1, lh_lb + 1, rh_lb, rh_ub); r.union_ (tmp); - if (wi::eq_p (lh_range, 3)) + if (lh_range == 3) { wi_fold (tmp, type, lh_lb + 2, lh_lb + 2, rh_lb, rh_ub); r.union_ (tmp); --- gcc/testsuite/g++.dg/opt/pr104334.C.jj 2022-02-02 14:35:51.184657968 +0100 +++ gcc/testsuite/g++.dg/opt/pr104334.C 2022-02-02 14:37:14.888478594 +0100 @@ -0,0 +1,40 @@ +// PR tree-optimization/104334 +// { dg-do run { target c++11 } } +// { dg-options "-O2 --param logical-op-non-short-circuit=0" } + +enum class A { A0, A1, A2, A3 }; +int x; + +__attribute__((noipa)) void +baz () +{ + x = 1; +} + +struct B { + unsigned b : 2; + + A + foo () const + { + return static_cast (b); + } + + __attribute__((noinline)) void + bar () + { + if (foo () == A::A2 || foo () == A::A3) + baz (); + } +}; + +int +main () +{ + B c; + c.b = 2; + c.bar (); + if (x != 1) + __builtin_abort (); + return 0; +} Jakub