From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.129.124]) by sourceware.org (Postfix) with ESMTPS id A5278385AC1F for ; Sun, 6 Nov 2022 16:14:34 +0000 (GMT) DMARC-Filter: OpenDMARC Filter v1.4.1 sourceware.org A5278385AC1F Authentication-Results: sourceware.org; dmarc=pass (p=none dis=none) header.from=redhat.com Authentication-Results: sourceware.org; spf=pass smtp.mailfrom=redhat.com DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1667751273; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding; bh=5e/RYt1uMG+KDgPN1jYlFakKuvb3AbcrrlgqVM3eMzw=; b=b45tUJCsAwtynZpRvxev9bkiQQGC92oxwYK2hRFQ827k0ujgT2TkmPvtx9Gf78+81NfSGu HC8CiLyQ2OwfiGupVZ5sfRd/ej0jC/CajOernd4LDOrZUVZUvxgWsy6mvsATlGrUiiTwa4 3SkI6RnAHyJGYVSAZ3mxR8f0uoMtKEU= Received: from mimecast-mx02.redhat.com (mx3-rdu2.redhat.com [66.187.233.73]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-625-tDBU4uyqNnCKWJHAb4qYMw-1; Sun, 06 Nov 2022 11:14:30 -0500 X-MC-Unique: tDBU4uyqNnCKWJHAb4qYMw-1 Received: from smtp.corp.redhat.com (int-mx10.intmail.prod.int.rdu2.redhat.com [10.11.54.10]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id C921A3803921; Sun, 6 Nov 2022 16:14:29 +0000 (UTC) Received: from abulafia.quesejoda.com (unknown [10.39.192.2]) by smtp.corp.redhat.com (Postfix) with ESMTPS id 72BCB492B0B; Sun, 6 Nov 2022 16:14:29 +0000 (UTC) Received: from abulafia.quesejoda.com (localhost [127.0.0.1]) by abulafia.quesejoda.com (8.17.1/8.17.1) with ESMTPS id 2A6GEQ4V522617 (version=TLSv1.3 cipher=TLS_AES_256_GCM_SHA384 bits=256 verify=NOT); Sun, 6 Nov 2022 17:14:27 +0100 Received: (from aldyh@localhost) by abulafia.quesejoda.com (8.17.1/8.17.1/Submit) id 2A6GEQHY522616; Sun, 6 Nov 2022 17:14:26 +0100 From: Aldy Hernandez To: GCC patches Cc: Richard Biener , Jakub Jelinek , Andrew MacLeod , Aldy Hernandez Subject: [PATCH] Use bit-CCP in range-ops. Date: Sun, 6 Nov 2022 17:14:20 +0100 Message-Id: <20221106161420.522485-1-aldyh@redhat.com> MIME-Version: 1.0 X-Scanned-By: MIMEDefang 3.1 on 10.11.54.10 X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Transfer-Encoding: 8bit Content-Type: text/plain; charset="US-ASCII"; x-default=true X-Spam-Status: No, score=-11.8 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH,DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,GIT_PATCH_0,RCVD_IN_DNSWL_NONE,RCVD_IN_MSPIKE_H2,SPF_HELO_NONE,SPF_NONE,TXREP autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on server2.sourceware.org List-Id: After Jakub and Richi's suggestion of using the same representation for tracking known bits as we do in CCP, I took a peek at the code and realized there's a plethora of bit-tracking code there that we could be sharing with range-ops. For example, the multiplication optimizations are way better than what I had cobbled together. For that matter, our maybe nonzero tracking as a whole has a lot of room for improvement. Being the lazy ass that I am, I think we should just use one code base (CCP's). This patch provides a thin wrapper for converting the irange maybe nonzero bits to what CCP requires, and uses that to call into bit_value_binop(). I have so far converted the MULT_EXPR range-op entry to use it, as the DIV_EXPR entry we have gets a case CCP doesn't get so I'd like to contribute the enhancement to CCP before converting over. I'd like to use this approach with the dozen or so tree_code's that are handled in CCP, thus saving us from having to implement any of them :). Early next season I'd like to change irange's internal representation to a pair of value / mask, and start tracking all known bits. This ties in nicely with our plan for tracking known set bits. Perhaps if the stars align, we could merge the bit twiddling in CCP into range-ops and have a central repository for it. That is, once we make the switch to wide-ints, and assuming there are no performance issues. Note that range-ops is our lowest level abstraction. i.e. it's just the math, there's no GORI or ranger, or even the concept of a symbolic or SSA. I'd love to hear comments and ideas, and if no one objects push this. Please let me know if I missed anything. Tested on x86-64 Linux. gcc/ChangeLog: * range-op.cc (irange_to_masked_value): New. (update_known_bitmask): New. (operator_mult::fold_range): Call update_known_bitmask. --- gcc/range-op.cc | 63 +++++++++++++++++++++++++++++++++++++++---------- 1 file changed, 50 insertions(+), 13 deletions(-) diff --git a/gcc/range-op.cc b/gcc/range-op.cc index 25c004d8287..6d9914d8d12 100644 --- a/gcc/range-op.cc +++ b/gcc/range-op.cc @@ -46,6 +46,54 @@ along with GCC; see the file COPYING3. If not see #include "wide-int.h" #include "value-relation.h" #include "range-op.h" +#include "tree-ssa-ccp.h" + +// Convert irange bitmasks into a VALUE MASK pair suitable for calling CCP. + +static void +irange_to_masked_value (const irange &r, widest_int &value, widest_int &mask) +{ + if (r.singleton_p ()) + { + mask = 0; + value = widest_int::from (r.lower_bound (), TYPE_SIGN (r.type ())); + } + else + { + mask = widest_int::from (r.get_nonzero_bits (), TYPE_SIGN (r.type ())); + value = 0; + } +} + +// Update the known bitmasks in R when applying the operation CODE to +// LH and RH. + +static void +update_known_bitmask (irange &r, tree_code code, + const irange &lh, const irange &rh) +{ + if (r.undefined_p ()) + return; + + widest_int value, mask, lh_mask, rh_mask, lh_value, rh_value; + tree type = r.type (); + signop sign = TYPE_SIGN (type); + int prec = TYPE_PRECISION (type); + signop lh_sign = TYPE_SIGN (lh.type ()); + signop rh_sign = TYPE_SIGN (rh.type ()); + int lh_prec = TYPE_PRECISION (lh.type ()); + int rh_prec = TYPE_PRECISION (rh.type ()); + + irange_to_masked_value (lh, lh_value, lh_mask); + irange_to_masked_value (rh, rh_value, rh_mask); + bit_value_binop (code, sign, prec, &value, &mask, + lh_sign, lh_prec, lh_value, lh_mask, + rh_sign, rh_prec, rh_value, rh_mask); + + int_range<2> tmp (type); + tmp.set_nonzero_bits (value | mask); + r.intersect (tmp); +} // Return the upper limit for a type. @@ -1774,21 +1822,10 @@ operator_mult::fold_range (irange &r, tree type, if (!cross_product_operator::fold_range (r, type, lh, rh, trio)) return false; - if (lh.undefined_p ()) + if (lh.undefined_p () || rh.undefined_p ()) return true; - tree t; - if (rh.singleton_p (&t)) - { - wide_int w = wi::to_wide (t); - int shift = wi::exact_log2 (w); - if (shift != -1) - { - wide_int nz = lh.get_nonzero_bits (); - nz = wi::lshift (nz, shift); - r.set_nonzero_bits (nz); - } - } + update_known_bitmask (r, MULT_EXPR, lh, rh); return true; } -- 2.38.1