@@ -120,6 +120,7 @@ along with GCC; see the file COPYING3. If not see
#include "params.h"
#include "ipa-inline.h"
#include "ipa-utils.h"
+#include "tree-ssa-ccp.h"
template <typename valtype> class ipcp_value;
@@ -266,6 +267,40 @@ private:
bool meet_with_1 (unsigned new_align, unsigned new_misalign);
};
+/* Lattice of known bits, only capable of holding one value.
+ Similar to ccp_prop_value_t, mask represents which bits of value are constant.
+ If a bit in mask is set to 0, then the corresponding bit in
+ value is known to be constant. */
+
+class ipcp_bits_lattice
+{
+public:
+ bool bottom_p () { return lattice_val == IPA_BITS_VARYING; }
+ bool top_p () { return lattice_val == IPA_BITS_UNDEFINED; }
+ bool constant_p () { return lattice_val == IPA_BITS_CONSTANT; }
+ bool set_to_bottom ();
+ bool set_to_constant (widest_int, widest_int, signop, unsigned);
+
+ widest_int get_value () { return value; }
+ widest_int get_mask () { return mask; }
+ signop get_sign () { return sgn; }
+ unsigned get_precision () { return precision; }
+
+ bool meet_with (ipcp_bits_lattice& other, enum tree_code, tree);
+ bool meet_with (widest_int, widest_int, signop, unsigned);
+
+ void print (FILE *);
+
+private:
+ enum { IPA_BITS_UNDEFINED, IPA_BITS_CONSTANT, IPA_BITS_VARYING } lattice_val;
+ widest_int value, mask;
+ signop sgn;
+ unsigned precision;
+
+ bool meet_with_1 (widest_int, widest_int);
+ void get_value_and_mask (tree, widest_int *, widest_int *);
+};
+
/* Structure containing lattices for a parameter itself and for pieces of
aggregates that are passed in the parameter or by a reference in a parameter
plus some other useful flags. */
@@ -281,6 +316,8 @@ public:
ipcp_agg_lattice *aggs;
/* Lattice describing known alignment. */
ipcp_alignment_lattice alignment;
+ /* Lattice describing known bits. */
+ ipcp_bits_lattice bits_lattice;
/* Number of aggregate lattices */
int aggs_count;
/* True if aggregate data were passed by reference (as opposed to by
@@ -458,6 +495,21 @@ ipcp_alignment_lattice::print (FILE * f)
fprintf (f, " Alignment %u, misalignment %u\n", align, misalign);
}
+void
+ipcp_bits_lattice::print (FILE *f)
+{
+ if (top_p ())
+ fprintf (f, " Bits unknown (TOP)\n");
+ else if (bottom_p ())
+ fprintf (f, " Bits unusable (BOTTOM)\n");
+ else
+ {
+ fprintf (f, " Bits: value = "); print_hex (get_value (), f);
+ fprintf (f, ", mask = "); print_hex (get_mask (), f);
+ fprintf (f, "\n");
+ }
+}
+
/* Print all ipcp_lattices of all functions to F. */
static void
@@ -484,6 +536,7 @@ print_all_lattices (FILE * f, bool dump_sources, bool dump_benefits)
fprintf (f, " ctxs: ");
plats->ctxlat.print (f, dump_sources, dump_benefits);
plats->alignment.print (f);
+ plats->bits_lattice.print (f);
if (plats->virt_call)
fprintf (f, " virt_call flag set\n");
@@ -911,6 +964,161 @@ ipcp_alignment_lattice::meet_with (const ipcp_alignment_lattice &other,
return meet_with_1 (other.align, adjusted_misalign);
}
+/* Set lattice value to bottom, if it already isn't the case. */
+
+bool
+ipcp_bits_lattice::set_to_bottom ()
+{
+ if (bottom_p ())
+ return false;
+ lattice_val = IPA_BITS_VARYING;
+ value = 0;
+ mask = -1;
+ return true;
+}
+
+/* Set to constant if it isn't already. Only meant to be called
+ when switching state from TOP. */
+
+bool
+ipcp_bits_lattice::set_to_constant (widest_int value, widest_int mask,
+ signop sgn, unsigned precision)
+{
+ gcc_assert (top_p ());
+ this->lattice_val = IPA_BITS_CONSTANT;
+ this->value = value;
+ this->mask = mask;
+ this->sgn = sgn;
+ this->precision = precision;
+ return true;
+}
+
+/* Convert operand to value, mask form. */
+
+void
+ipcp_bits_lattice::get_value_and_mask (tree operand, widest_int *valuep, widest_int *maskp)
+{
+ wide_int get_nonzero_bits (const_tree);
+
+ if (TREE_CODE (operand) == INTEGER_CST)
+ {
+ *valuep = wi::to_widest (operand);
+ *maskp = 0;
+ }
+ else if (TREE_CODE (operand) == SSA_NAME)
+ {
+ *valuep = 0;
+ *maskp = widest_int::from (get_nonzero_bits (operand), UNSIGNED);
+ }
+ else
+ gcc_unreachable ();
+}
+
+/* Meet operation, similar to ccp_lattice_meet, we xor values
+ if this->value, value have different values at same bit positions, we want
+ to drop that bit to varying. Return true if mask is changed.
+ This function assumes that the lattice value is in CONSTANT state */
+
+bool
+ipcp_bits_lattice::meet_with_1 (widest_int value, widest_int mask)
+{
+ gcc_assert (constant_p ());
+
+ widest_int old_mask = this->mask;
+ this->mask = (this->mask | mask) | (this->value ^ value);
+
+ if (wi::sext (this->mask, this->precision) == -1)
+ return set_to_bottom ();
+
+ bool changed = this->mask != old_mask;
+ return changed;
+}
+
+/* Meet the bits lattice with operand
+ described by <value, mask, sgn, precision. */
+
+bool
+ipcp_bits_lattice::meet_with (widest_int value, widest_int mask,
+ signop sgn, unsigned precision)
+{
+ if (bottom_p ())
+ return false;
+
+ if (top_p ())
+ {
+ if (wi::sext (mask, precision) == -1)
+ return set_to_bottom ();
+ return set_to_constant (value, mask, sgn, precision);
+ }
+
+ return meet_with_1 (value, mask);
+}
+
+/* Meet bits lattice with the result of bit_value_binop_1 (other, operand)
+ if code is binary operation or bit_value_unop_1 (other) if code is unary op.
+ In the case when code is nop_expr, no adjustment is required. */
+
+bool
+ipcp_bits_lattice::meet_with (ipcp_bits_lattice& other, enum tree_code code, tree operand)
+{
+ if (other.bottom_p ())
+ return set_to_bottom ();
+
+ if (bottom_p () || other.top_p ())
+ return false;
+
+ widest_int adjusted_value, adjusted_mask;
+
+ if (TREE_CODE_CLASS (code) == tcc_binary)
+ {
+ tree type = TREE_TYPE (operand);
+ gcc_assert (INTEGRAL_TYPE_P (type));
+ widest_int o_value, o_mask;
+ get_value_and_mask (operand, &o_value, &o_mask);
+
+ signop sgn = other.get_sign ();
+ unsigned prec = other.get_precision ();
+
+ bit_value_binop_1 (code, sgn, prec, &adjusted_value, &adjusted_mask,
+ sgn, prec, other.get_value (), other.get_mask (),
+ TYPE_SIGN (type), TYPE_PRECISION (type), o_value, o_mask);
+
+ if (wi::sext (adjusted_mask, prec) == -1)
+ return set_to_bottom ();
+ }
+
+ else if (TREE_CODE_CLASS (code) == tcc_unary)
+ {
+ signop sgn = other.get_sign ();
+ unsigned prec = other.get_precision ();
+
+ bit_value_unop_1 (code, sgn, prec, &adjusted_value,
+ &adjusted_mask, sgn, prec, other.get_value (),
+ other.get_mask ());
+
+ if (wi::sext (adjusted_mask, prec) == -1)
+ return set_to_bottom ();
+ }
+
+ else if (code == NOP_EXPR)
+ {
+ adjusted_value = other.value;
+ adjusted_mask = other.mask;
+ }
+
+ else
+ return set_to_bottom ();
+
+ if (top_p ())
+ {
+ if (wi::sext (adjusted_mask, other.get_precision ()) == -1)
+ return set_to_bottom ();
+ return set_to_constant (adjusted_value, adjusted_mask, other.get_sign (), other.get_precision ());
+ }
+ else
+ return meet_with_1 (adjusted_value, adjusted_mask);
+}
+
/* Mark bot aggregate and scalar lattices as containing an unknown variable,
return true is any of them has not been marked as such so far. */
@@ -922,6 +1130,7 @@ set_all_contains_variable (struct ipcp_param_lattices *plats)
ret |= plats->ctxlat.set_contains_variable ();
ret |= set_agg_lats_contain_variable (plats);
ret |= plats->alignment.set_to_bottom ();
+ ret |= plats->bits_lattice.set_to_bottom ();
return ret;
}
@@ -1003,6 +1212,7 @@ initialize_node_lattices (struct cgraph_node *node)
plats->ctxlat.set_to_bottom ();
set_agg_lats_to_bottom (plats);
plats->alignment.set_to_bottom ();
+ plats->bits_lattice.set_to_bottom ();
}
else
set_all_contains_variable (plats);
@@ -1621,6 +1831,57 @@ propagate_alignment_accross_jump_function (cgraph_edge *cs,
}
}
+/* Propagate bits across jfunc that is associated with
+ edge cs and update dest_lattice accordingly. */
+
+bool
+propagate_bits_accross_jump_function (cgraph_edge *cs, ipa_jump_func *jfunc,
+ ipcp_bits_lattice *dest_lattice)
+{
+ if (dest_lattice->bottom_p ())
+ return false;
+
+ if (jfunc->type == IPA_JF_PASS_THROUGH)
+ {
+ struct ipa_node_params *caller_info = IPA_NODE_REF (cs->caller);
+ enum tree_code code = ipa_get_jf_pass_through_operation (jfunc);
+ tree operand = NULL_TREE;
+
+ if (code != NOP_EXPR)
+ operand = ipa_get_jf_pass_through_operand (jfunc);
+
+ int src_idx = ipa_get_jf_pass_through_formal_id (jfunc);
+ struct ipcp_param_lattices *src_lats
+ = ipa_get_parm_lattices (caller_info, src_idx);
+
+ /* Try to proapgate bits if src_lattice is bottom, but jfunc is known.
+ for eg consider:
+ int f(int x)
+ {
+ g (x & 0xff);
+ }
+ Assume lattice for x is bottom, however we can still propagate
+ result of x & 0xff == 0xff, which gets computed during ccp1 pass
+ and we store it in jump function during analysis stage. */
+
+ if (src_lats->bits_lattice.bottom_p ()
+ && jfunc->bits.known)
+ return dest_lattice->meet_with (jfunc->bits.value, jfunc->bits.mask,
+ jfunc->bits.sgn, jfunc->bits.precision);
+ else
+ return dest_lattice->meet_with (src_lats->bits_lattice, code, operand);
+ }
+
+ else if (jfunc->type == IPA_JF_ANCESTOR)
+ return dest_lattice->set_to_bottom ();
+
+ else if (jfunc->bits.known)
+ return dest_lattice->meet_with (jfunc->bits.value, jfunc->bits.mask,
+ jfunc->bits.sgn, jfunc->bits.precision);
+ else
+ return dest_lattice->set_to_bottom ();
+}
+
/* If DEST_PLATS already has aggregate items, check that aggs_by_ref matches
NEW_AGGS_BY_REF and if not, mark all aggs as bottoms and return true (in all
other cases, return false). If there are no aggregate items, set
@@ -1968,6 +2229,8 @@ propagate_constants_accross_call (struct cgraph_edge *cs)
&dest_plats->ctxlat);
ret |= propagate_alignment_accross_jump_function (cs, jump_func,
&dest_plats->alignment);
+ ret |= propagate_bits_accross_jump_function (cs, jump_func,
+ &dest_plats->bits_lattice);
ret |= propagate_aggs_accross_jump_function (cs, jump_func,
dest_plats);
}
@@ -4592,6 +4855,74 @@ ipcp_store_alignment_results (void)
}
}
+/* Look up all the bits information that we have discovered and copy it over
+ to the transformation summary. */
+
+static void
+ipcp_store_bits_results (void)
+{
+ cgraph_node *node;
+
+ FOR_EACH_FUNCTION_WITH_GIMPLE_BODY (node)
+ {
+ ipa_node_params *info = IPA_NODE_REF (node);
+ bool dumped_sth = false;
+ bool found_useful_result = false;
+
+ if (info->ipcp_orig_node)
+ info = IPA_NODE_REF (info->ipcp_orig_node);
+
+ unsigned count = ipa_get_param_count (info);
+ for (unsigned i = 0; i < count; i++)
+ {
+ ipcp_param_lattices *plats = ipa_get_parm_lattices (info, i);
+ if (plats->bits_lattice.constant_p ())
+ {
+ found_useful_result = true;
+ break;
+ }
+ }
+
+ if (!found_useful_result)
+ continue;
+
+ ipcp_grow_transformations_if_necessary ();
+ ipcp_transformation_summary *ts = ipcp_get_transformation_summary (node);
+ vec_safe_reserve_exact (ts->bits, count);
+
+ for (unsigned i = 0; i < count; i++)
+ {
+ ipcp_param_lattices *plats = ipa_get_parm_lattices (info, i);
+ ipa_bits bits_jfunc;
+
+ if (plats->bits_lattice.constant_p ())
+ {
+ bits_jfunc.known = true;
+ bits_jfunc.value = plats->bits_lattice.get_value ();
+ bits_jfunc.mask = plats->bits_lattice.get_mask ();
+ bits_jfunc.sgn = plats->bits_lattice.get_sign ();
+ bits_jfunc.precision = plats->bits_lattice.get_precision ();
+ }
+ else
+ bits_jfunc.known = false;
+
+ ts->bits->quick_push (bits_jfunc);
+ if (!dump_file || !bits_jfunc.known)
+ continue;
+ if (!dumped_sth)
+ {
+ fprintf (dump_file, "Propagated bits info for function %s/%i:\n",
+ node->name (), node->order);
+ dumped_sth = true;
+ }
+ fprintf (dump_file, " param %i: value = ", i);
+ print_hex (bits_jfunc.value, dump_file);
+ fprintf (dump_file, ", mask = ");
+ print_hex (bits_jfunc.mask, dump_file);
+ fprintf (dump_file, "\n");
+ }
+ }
+}
/* The IPCP driver. */
static unsigned int
@@ -4625,6 +4956,8 @@ ipcp_driver (void)
ipcp_decision_stage (&topo);
/* Store results of alignment propagation. */
ipcp_store_alignment_results ();
+ /* Store results of bits propagation. */
+ ipcp_store_bits_results ();
/* Free all IPCP structures. */
free_toporder_info (&topo);
@@ -302,6 +302,15 @@ ipa_print_node_jump_functions_for_edge (FILE *f, struct cgraph_edge *cs)
}
else
fprintf (f, " Unknown alignment\n");
+
+ if (jump_func->bits.known)
+ {
+ fprintf (f, " value: "); print_hex (jump_func->bits.value, f);
+ fprintf (f, ", mask: "); print_hex (jump_func->bits.mask, f);
+ fprintf (f, "\n");
+ }
+ else
+ fprintf (f, " Unknown bits\n");
}
}
@@ -381,6 +390,7 @@ ipa_set_jf_unknown (struct ipa_jump_func *jfunc)
{
jfunc->type = IPA_JF_UNKNOWN;
jfunc->alignment.known = false;
+ jfunc->bits.known = false;
}
/* Set JFUNC to be a copy of another jmp (to be used by jump function
@@ -1674,6 +1684,27 @@ ipa_compute_jump_functions_for_edge (struct ipa_func_body_info *fbi,
else
gcc_assert (!jfunc->alignment.known);
+ if (INTEGRAL_TYPE_P (TREE_TYPE (arg))
+ && (TREE_CODE (arg) == SSA_NAME || TREE_CODE (arg) == INTEGER_CST))
+ {
+ jfunc->bits.known = true;
+ jfunc->bits.sgn = TYPE_SIGN (TREE_TYPE (arg));
+ jfunc->bits.precision = TYPE_PRECISION (TREE_TYPE (arg));
+
+ if (TREE_CODE (arg) == SSA_NAME)
+ {
+ jfunc->bits.value = 0;
+ jfunc->bits.mask = widest_int::from (get_nonzero_bits (arg), UNSIGNED);
+ }
+ else
+ {
+ jfunc->bits.value = wi::to_widest (arg);
+ jfunc->bits.mask = 0;
+ }
+ }
+ else
+ gcc_assert (!jfunc->bits.known);
+
if (is_gimple_ip_invariant (arg)
|| (TREE_CODE (arg) == VAR_DECL
&& is_global_var (arg)
@@ -3690,6 +3721,18 @@ ipa_node_params_t::duplicate(cgraph_node *src, cgraph_node *dst,
for (unsigned i = 0; i < src_alignments->length (); ++i)
dst_alignments->quick_push ((*src_alignments)[i]);
}
+
+ if (src_trans && vec_safe_length (src_trans->bits) > 0)
+ {
+ ipcp_grow_transformations_if_necessary ();
+ src_trans = ipcp_get_transformation_summary (src);
+ const vec<ipa_bits, va_gc> *src_bits = src_trans->bits;
+ vec<ipa_bits, va_gc> *&dst_bits
+ = ipcp_get_transformation_summary (dst)->bits;
+ vec_safe_reserve_exact (dst_bits, src_bits->length ());
+ for (unsigned i = 0; i < src_bits->length (); ++i)
+ dst_bits->quick_push ((*src_bits)[i]);
+ }
}
/* Register our cgraph hooks if they are not already there. */
@@ -4609,6 +4652,17 @@ ipa_write_jump_function (struct output_block *ob,
streamer_write_uhwi (ob, jump_func->alignment.align);
streamer_write_uhwi (ob, jump_func->alignment.misalign);
}
+
+ bp = bitpack_create (ob->main_stream);
+ bp_pack_value (&bp, jump_func->bits.known, 1);
+ streamer_write_bitpack (&bp);
+ if (jump_func->bits.known)
+ {
+ streamer_write_wi (ob, jump_func->bits.value);
+ streamer_write_wi (ob, jump_func->bits.mask);
+ streamer_write_enum (ob->main_stream, signop, UNSIGNED + 1, jump_func->bits.sgn);
+ streamer_write_uhwi (ob, jump_func->bits.precision);
+ }
}
/* Read in jump function JUMP_FUNC from IB. */
@@ -4685,6 +4739,19 @@ ipa_read_jump_function (struct lto_input_block *ib,
}
else
jump_func->alignment.known = false;
+
+ bp = streamer_read_bitpack (ib);
+ bool bits_known = bp_unpack_value (&bp, 1);
+ if (bits_known)
+ {
+ jump_func->bits.known = true;
+ jump_func->bits.value = streamer_read_wi (ib);
+ jump_func->bits.mask = streamer_read_wi (ib);
+ jump_func->bits.sgn = streamer_read_enum (ib, signop, UNSIGNED + 1);
+ jump_func->bits.precision = streamer_read_uhwi (ib);
+ }
+ else
+ jump_func->bits.known = false;
}
/* Stream out parts of cgraph_indirect_call_info corresponding to CS that are
@@ -5050,6 +5117,31 @@ write_ipcp_transformation_info (output_block *ob, cgraph_node *node)
}
else
streamer_write_uhwi (ob, 0);
+
+ ts = ipcp_get_transformation_summary (node);
+ if (ts && vec_safe_length (ts->bits) > 0)
+ {
+ count = ts->bits->length ();
+ streamer_write_uhwi (ob, count);
+
+ for (unsigned i = 0; i < count; ++i)
+ {
+ const ipa_bits& bits_jfunc = (*ts->bits)[i];
+ struct bitpack_d bp = bitpack_create (ob->main_stream);
+ bp_pack_value (&bp, bits_jfunc.known, 1);
+ streamer_write_bitpack (&bp);
+ if (bits_jfunc.known)
+ {
+ streamer_write_wi (ob, bits_jfunc.value);
+ streamer_write_wi (ob, bits_jfunc.mask);
+ streamer_write_enum (ob->main_stream, signop,
+ UNSIGNED + 1, bits_jfunc.sgn);
+ streamer_write_uhwi (ob, bits_jfunc.precision);
+ }
+ }
+ }
+ else
+ streamer_write_uhwi (ob, 0);
}
/* Stream in the aggregate value replacement chain for NODE from IB. */
@@ -5102,6 +5194,28 @@ read_ipcp_transformation_info (lto_input_block *ib, cgraph_node *node,
}
}
}
+
+ count = streamer_read_uhwi (ib);
+ if (count > 0)
+ {
+ ipcp_grow_transformations_if_necessary ();
+ ipcp_transformation_summary *ts = ipcp_get_transformation_summary (node);
+ vec_safe_grow_cleared (ts->bits, count);
+
+ for (i = 0; i < count; i++)
+ {
+ ipa_bits& bits_jfunc = (*ts->bits)[i];
+ struct bitpack_d bp = streamer_read_bitpack (ib);
+ bits_jfunc.known = bp_unpack_value (&bp, 1);
+ if (bits_jfunc.known)
+ {
+ bits_jfunc.value = streamer_read_wi (ib);
+ bits_jfunc.mask = streamer_read_wi (ib);
+ bits_jfunc.sgn = streamer_read_enum (ib, signop, UNSIGNED + 1);
+ bits_jfunc.precision = streamer_read_uhwi (ib);
+ }
+ }
+ }
}
/* Write all aggregate replacement for nodes in set. */
@@ -5404,6 +5518,55 @@ ipcp_update_alignments (struct cgraph_node *node)
}
}
+/* Update bits info of formal parameters as described in
+ ipcp_transformation_summary. */
+
+static void
+ipcp_update_bits (struct cgraph_node *node)
+{
+ tree parm = DECL_ARGUMENTS (node->decl);
+ tree next_parm = parm;
+ ipcp_transformation_summary *ts = ipcp_get_transformation_summary (node);
+
+ if (!ts || vec_safe_length (ts->bits) == 0)
+ return;
+
+ vec<ipa_bits, va_gc> &bits = *ts->bits;
+ unsigned count = bits.length ();
+
+ for (unsigned i = 0; i < count; ++i, parm = next_parm)
+ {
+ if (node->clone.combined_args_to_skip
+ && bitmap_bit_p (node->clone.combined_args_to_skip, i))
+ continue;
+
+ gcc_checking_assert (parm);
+ next_parm = DECL_CHAIN (parm);
+
+ if (!bits[i].known
+ || !INTEGRAL_TYPE_P (TREE_TYPE (parm))
+ || !is_gimple_reg (parm))
+ continue;
+
+ tree ddef = ssa_default_def (DECL_STRUCT_FUNCTION (node->decl), parm);
+ if (!ddef)
+ continue;
+
+ if (dump_file)
+ {
+ fprintf (dump_file, "Adjusting mask for param %u to ", i);
+ print_hex (bits[i].mask, dump_file);
+ fprintf (dump_file, "\n");
+ }
+
+ unsigned prec = TYPE_PRECISION (TREE_TYPE (ddef));
+ wide_int nonzero_bits = wide_int::from (bits[i].mask, prec, UNSIGNED)
+ | wide_int::from (bits[i].value, prec, bits[i].sgn);
+ set_nonzero_bits (ddef, nonzero_bits);
+ DECL_SET_BY_IPA (parm) = 1;
+ }
+}
+
/* IPCP transformation phase doing propagation of aggregate values. */
unsigned int
@@ -5423,6 +5586,7 @@ ipcp_transform_function (struct cgraph_node *node)
node->name (), node->order);
ipcp_update_alignments (node);
+ ipcp_update_bits (node);
aggval = ipa_get_agg_replacements_for_node (node);
if (!aggval)
return 0;
@@ -154,6 +154,16 @@ struct GTY(()) ipa_alignment
unsigned misalign;
};
+/* Information about zero/non-zero bits. */
+struct GTY(()) ipa_bits
+{
+ bool known;
+ widest_int value;
+ widest_int mask;
+ enum signop sgn;
+ unsigned precision;
+};
+
/* A jump function for a callsite represents the values passed as actual
arguments of the callsite. See enum jump_func_type for the various
types of jump functions supported. */
@@ -166,6 +176,9 @@ struct GTY (()) ipa_jump_func
/* Information about alignment of pointers. */
struct ipa_alignment alignment;
+ /* Information about zero/non-zero bits. */
+ struct ipa_bits bits;
+
enum jump_func_type type;
/* Represents a value of a jump function. pass_through is used only in jump
function context. constant represents the actual constant in constant jump
@@ -482,6 +495,8 @@ struct GTY(()) ipcp_transformation_summary
ipa_agg_replacement_value *agg_values;
/* Alignment information for pointers. */
vec<ipa_alignment, va_gc> *alignments;
+ /* Known bits information. */
+ vec<ipa_bits, va_gc> *bits;
};
void ipa_set_node_agg_value_chain (struct cgraph_node *node,
@@ -712,7 +712,7 @@ make_new_block (struct function *fn, unsigned int index)
/* Read a wide-int. */
-static widest_int
+widest_int
streamer_read_wi (struct lto_input_block *ib)
{
HOST_WIDE_INT a[WIDE_INT_MAX_ELTS];
@@ -1830,7 +1830,7 @@ output_ssa_names (struct output_block *ob, struct function *fn)
/* Output a wide-int. */
-static void
+void
streamer_write_wi (struct output_block *ob,
const widest_int &w)
{
@@ -1225,4 +1225,7 @@ DEFINE_DECL_STREAM_FUNCS (TYPE_DECL, type_decl)
DEFINE_DECL_STREAM_FUNCS (NAMESPACE_DECL, namespace_decl)
DEFINE_DECL_STREAM_FUNCS (LABEL_DECL, label_decl)
+widest_int streamer_read_wi (struct lto_input_block *);
+void streamer_write_wi (struct output_block *, const widest_int &);
+
#endif /* GCC_LTO_STREAMER_H */
new file mode 100644
@@ -0,0 +1,33 @@
+/* Propagate 0xff from main to f3 to f2. */
+
+/* { dg-do compile } */
+/* { dg-options "-O2 -fno-early-inlining -fdump-ipa-cp -fdump-tree-optimized" } */
+
+int pass_test(void);
+int fail_test(void);
+
+__attribute__((noinline, noclone))
+static int f2(int x)
+{
+ if (x > 300)
+ return fail_test();
+ else
+ return pass_test();
+}
+
+__attribute__((noinline, noclone))
+static int f3(int y)
+{
+ int k = f2(y);
+ return k;
+}
+
+int main(int argc)
+{
+ int k = argc & 0xff;
+ int a = f3(k);
+ return a;
+}
+
+/* { dg-final { scan-ipa-dump-times "Adjusting mask for param 0 to 0xff" 2 "cp" } } */
+/* { dg-final { scan-tree-dump-not "fail_test" "optimized" } } */
new file mode 100644
@@ -0,0 +1,37 @@
+/* x's mask should be meet(0xc, 0x3) == 0xf */
+
+/* { dg-do compile } */
+/* { dg-options "-O2 -fno-early-inlining -fdump-ipa-cp" } */
+
+__attribute__((noinline))
+static int f1(int x)
+{
+ if (x > 300)
+ return 1;
+ else
+ return 2;
+}
+
+__attribute__((noinline))
+static int f2(int y)
+{
+ return f1(y & 0x03);
+}
+
+__attribute__((noinline))
+static int f3(int z)
+{
+ return f1(z & 0xc);
+}
+
+extern int a;
+extern int b;
+
+int main(void)
+{
+ int k = f2(a);
+ int l = f3(b);
+ return k + l;
+}
+
+/* { dg-final { scan-ipa-dump "Adjusting mask for param 0 to 0xf" "cp" } } */
new file mode 100644
@@ -0,0 +1,24 @@
+/* { dg-do compile } */
+/* { dg-options "-O2 -fno-early-inlining -fdump-ipa-cp" } */
+
+__attribute__((noinline))
+static int f(int x)
+{
+ int f2(int);
+
+ if (x > 300)
+ {
+ int z = f(x + 1);
+ return f2 (z);
+ }
+ else
+ return 2;
+}
+
+int main(int argc, char **argv)
+{
+ int k = f(argc & 0xff);
+ return k;
+}
+
+/* { dg-final { scan-ipa-dump-not "Adjusting mask for" "cp" } } */
new file mode 100644
@@ -0,0 +1,21 @@
+/* { dg-do compile } */
+/* { dg-options "-O2 -fno-early-inlining -fdump-ipa-cp" } */
+
+__attribute__((noinline))
+static int f(int x)
+{
+ if (x > 300)
+ return 1;
+ else
+ return 2;
+}
+
+int main(void)
+{
+ int a = f(1);
+ int b = f(2);
+ int c = f(4);
+ return a + b + c;
+}
+
+/* { dg-final { scan-ipa-dump "Adjusting mask for param 0 to 0x7" "cp" } } */
new file mode 100644
@@ -0,0 +1,26 @@
+/* { dg-do compile } */
+/* { dg-options "-O2 -fno-early-inlining -fdump-ipa-cp" } */
+
+__attribute__((noinline))
+static int f1(int x)
+{
+ if (x > 20)
+ return 1;
+ else
+ return 2;
+}
+
+__attribute__((noinline))
+static int f2(int y)
+{
+ return f1 (y & 0x3);
+}
+
+int main(int argc, char **argv)
+{
+ int z = f2 (argc & 0xff);
+ int k = f1 (argc & 0xc);
+ return z + k;
+}
+
+/* { dg-final { scan-ipa-dump "Adjusting mask for param 0 to 0xf" "cp" } } */
@@ -1558,7 +1558,8 @@ struct GTY(()) tree_decl_common {
/* DECL_ALIGN. It should have the same size as TYPE_ALIGN. */
unsigned int align : 6;
- /* 20 bits unused. */
+ unsigned set_by_ipa: 1;
+ /* 19 bits unused. */
/* UID for points-to sets, stable over copying from inlining. */
unsigned int pt_uid;
@@ -142,7 +142,7 @@ along with GCC; see the file COPYING3. If not see
#include "cfgloop.h"
#include "stor-layout.h"
#include "optabs-query.h"
-
+#include "tree-ssa-ccp.h"
/* Possible lattice values. */
typedef enum
@@ -287,7 +287,11 @@ get_default_value (tree var)
{
val.lattice_val = CONSTANT;
val.value = build_zero_cst (TREE_TYPE (var));
- val.mask = extend_mask (nonzero_bits);
+ if (SSA_NAME_VAR (var) && TREE_CODE (SSA_NAME_VAR (var)) == PARM_DECL
+ && DECL_SET_BY_IPA (SSA_NAME_VAR (var)))
+ val.mask = widest_int::from (nonzero_bits, TYPE_SIGN (TREE_TYPE (SSA_NAME_VAR (var))));
+ else
+ val.mask = extend_mask (nonzero_bits);
}
}
}
@@ -537,9 +541,9 @@ set_lattice_value (tree var, ccp_prop_value_t *new_val)
static ccp_prop_value_t get_value_for_expr (tree, bool);
static ccp_prop_value_t bit_value_binop (enum tree_code, tree, tree, tree);
-static void bit_value_binop_1 (enum tree_code, tree, widest_int *, widest_int *,
- tree, const widest_int &, const widest_int &,
- tree, const widest_int &, const widest_int &);
+void bit_value_binop_1 (enum tree_code, signop, unsigned, widest_int *, widest_int *,
+ signop, unsigned, const widest_int &, const widest_int &,
+ signop, unsigned, const widest_int &, const widest_int &);
/* Return a widest_int that can be used for bitwise simplifications
from VAL. */
@@ -895,7 +899,7 @@ do_dbg_cnt (void)
Return TRUE when something was optimized. */
static bool
-ccp_finalize (bool nonzero_p)
+ccp_finalize (bool nonzero_p ATTRIBUTE_UNUSED)
{
bool something_changed;
unsigned i;
@@ -913,10 +917,7 @@ ccp_finalize (bool nonzero_p)
if (!name
|| (!POINTER_TYPE_P (TREE_TYPE (name))
- && (!INTEGRAL_TYPE_P (TREE_TYPE (name))
- /* Don't record nonzero bits before IPA to avoid
- using too much memory. */
- || !nonzero_p)))
+ && (!INTEGRAL_TYPE_P (TREE_TYPE (name)))))
continue;
val = get_value (name);
@@ -1225,10 +1226,11 @@ ccp_fold (gimple *stmt)
RVAL and RMASK representing a value of type RTYPE and set
the value, mask pair *VAL and *MASK to the result. */
-static void
-bit_value_unop_1 (enum tree_code code, tree type,
+void
+bit_value_unop_1 (enum tree_code code, signop type_sgn, unsigned type_precision,
widest_int *val, widest_int *mask,
- tree rtype, const widest_int &rval, const widest_int &rmask)
+ signop rtype_sgn, unsigned rtype_precision,
+ const widest_int &rval, const widest_int &rmask)
{
switch (code)
{
@@ -1241,25 +1243,23 @@ bit_value_unop_1 (enum tree_code code, tree type,
{
widest_int temv, temm;
/* Return ~rval + 1. */
- bit_value_unop_1 (BIT_NOT_EXPR, type, &temv, &temm, type, rval, rmask);
- bit_value_binop_1 (PLUS_EXPR, type, val, mask,
- type, temv, temm, type, 1, 0);
+ bit_value_unop_1 (BIT_NOT_EXPR, type_sgn, type_precision, &temv, &temm,
+ type_sgn, type_precision, rval, rmask);
+ bit_value_binop_1 (PLUS_EXPR, type_sgn, type_precision, val, mask,
+ type_sgn, type_precision, temv, temm,
+ type_sgn, type_precision, 1, 0);
break;
}
CASE_CONVERT:
{
- signop sgn;
-
/* First extend mask and value according to the original type. */
- sgn = TYPE_SIGN (rtype);
- *mask = wi::ext (rmask, TYPE_PRECISION (rtype), sgn);
- *val = wi::ext (rval, TYPE_PRECISION (rtype), sgn);
+ *mask = wi::ext (rmask, rtype_precision, rtype_sgn);
+ *val = wi::ext (rval, rtype_precision, rtype_sgn);
/* Then extend mask and value according to the target type. */
- sgn = TYPE_SIGN (type);
- *mask = wi::ext (*mask, TYPE_PRECISION (type), sgn);
- *val = wi::ext (*val, TYPE_PRECISION (type), sgn);
+ *mask = wi::ext (*mask, type_precision, type_sgn);
+ *val = wi::ext (*val, type_precision, type_sgn);
break;
}
@@ -1273,15 +1273,16 @@ bit_value_unop_1 (enum tree_code code, tree type,
R1VAL, R1MASK and R2VAL, R2MASK representing a values of type R1TYPE
and R2TYPE and set the value, mask pair *VAL and *MASK to the result. */
-static void
-bit_value_binop_1 (enum tree_code code, tree type,
+void
+bit_value_binop_1 (enum tree_code code, signop type_sgn, unsigned type_precision,
widest_int *val, widest_int *mask,
- tree r1type, const widest_int &r1val,
- const widest_int &r1mask, tree r2type,
+ signop r1type_sgn, unsigned r1type_precision,
+ const widest_int &r1val, const widest_int &r1mask,
+ signop r2type_sgn, unsigned r2type_precision,
const widest_int &r2val, const widest_int &r2mask)
{
- signop sgn = TYPE_SIGN (type);
- int width = TYPE_PRECISION (type);
+ signop sgn = type_sgn;
+ int width = (int) type_precision;
bool swap_p = false;
/* Assume we'll get a constant result. Use an initial non varying
@@ -1407,11 +1408,11 @@ bit_value_binop_1 (enum tree_code code, tree type,
case MINUS_EXPR:
{
widest_int temv, temm;
- bit_value_unop_1 (NEGATE_EXPR, r2type, &temv, &temm,
- r2type, r2val, r2mask);
- bit_value_binop_1 (PLUS_EXPR, type, val, mask,
- r1type, r1val, r1mask,
- r2type, temv, temm);
+ bit_value_unop_1 (NEGATE_EXPR, r2type_sgn, r2type_precision, &temv, &temm,
+ r2type_sgn, r2type_precision, r2val, r2mask);
+ bit_value_binop_1 (PLUS_EXPR, type_sgn, type_precision, val, mask,
+ r1type_sgn, r1type_precision, r1val, r1mask,
+ r2type_sgn, r2type_precision, temv, temm);
break;
}
@@ -1473,7 +1474,7 @@ bit_value_binop_1 (enum tree_code code, tree type,
break;
/* For comparisons the signedness is in the comparison operands. */
- sgn = TYPE_SIGN (r1type);
+ sgn = r1type_sgn;
/* If we know the most significant bits we know the values
value ranges by means of treating varying bits as zero
@@ -1526,8 +1527,9 @@ bit_value_unop (enum tree_code code, tree type, tree rhs)
gcc_assert ((rval.lattice_val == CONSTANT
&& TREE_CODE (rval.value) == INTEGER_CST)
|| wi::sext (rval.mask, TYPE_PRECISION (TREE_TYPE (rhs))) == -1);
- bit_value_unop_1 (code, type, &value, &mask,
- TREE_TYPE (rhs), value_to_wide_int (rval), rval.mask);
+ bit_value_unop_1 (code, TYPE_SIGN (type), TYPE_PRECISION (type), &value, &mask,
+ TYPE_SIGN (TREE_TYPE (rhs)), TYPE_PRECISION (TREE_TYPE (rhs)),
+ value_to_wide_int (rval), rval.mask);
if (wi::sext (mask, TYPE_PRECISION (type)) != -1)
{
val.lattice_val = CONSTANT;
@@ -1572,9 +1574,11 @@ bit_value_binop (enum tree_code code, tree type, tree rhs1, tree rhs2)
&& TREE_CODE (r2val.value) == INTEGER_CST)
|| wi::sext (r2val.mask,
TYPE_PRECISION (TREE_TYPE (rhs2))) == -1);
- bit_value_binop_1 (code, type, &value, &mask,
- TREE_TYPE (rhs1), value_to_wide_int (r1val), r1val.mask,
- TREE_TYPE (rhs2), value_to_wide_int (r2val), r2val.mask);
+ bit_value_binop_1 (code, TYPE_SIGN (type), TYPE_PRECISION (type), &value, &mask,
+ TYPE_SIGN (TREE_TYPE (rhs1)), TYPE_PRECISION (TREE_TYPE (rhs1)),
+ value_to_wide_int (r1val), r1val.mask,
+ TYPE_SIGN (TREE_TYPE (rhs2)), TYPE_PRECISION (TREE_TYPE (rhs2)),
+ value_to_wide_int (r2val), r2val.mask);
if (wi::sext (mask, TYPE_PRECISION (type)) != -1)
{
val.lattice_val = CONSTANT;
@@ -1673,9 +1677,9 @@ bit_value_assume_aligned (gimple *stmt, tree attr, ccp_prop_value_t ptrval,
align = build_int_cst_type (type, -aligni);
alignval = get_value_for_expr (align, true);
- bit_value_binop_1 (BIT_AND_EXPR, type, &value, &mask,
- type, value_to_wide_int (ptrval), ptrval.mask,
- type, value_to_wide_int (alignval), alignval.mask);
+ bit_value_binop_1 (BIT_AND_EXPR, TYPE_SIGN (type), TYPE_PRECISION (type), &value, &mask,
+ TYPE_SIGN (type), TYPE_PRECISION (type), value_to_wide_int (ptrval), ptrval.mask,
+ TYPE_SIGN (type), TYPE_PRECISION (type), value_to_wide_int (alignval), alignval.mask);
if (wi::sext (mask, TYPE_PRECISION (type)) != -1)
{
val.lattice_val = CONSTANT;
new file mode 100644
@@ -0,0 +1,30 @@
+/* Copyright (C) 2016-2016 Free Software Foundation, Inc.
+
+This file is part of GCC.
+
+GCC is free software; you can redistribute it and/or modify it
+under the terms of the GNU General Public License as published by the
+Free Software Foundation; either version 3, or (at your option) any
+later version.
+
+GCC is distributed in the hope that it will be useful, but WITHOUT
+ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
+for more details.
+
+You should have received a copy of the GNU General Public License
+along with GCC; see the file COPYING3. If not see
+<http://www.gnu.org/licenses/>. */
+
+#ifndef TREE_SSA_CCP_H
+#define TREE_SSA_CCP_H
+
+void bit_value_binop_1 (enum tree_code, signop, unsigned, widest_int *, widest_int *,
+ signop, unsigned, const widest_int &, const widest_int &,
+ signop, unsigned, const widest_int &, const widest_int &);
+
+void bit_value_unop_1 (enum tree_code, signop, unsigned, widest_int *, widest_int *,
+ signop, unsigned, const widest_int &, const widest_int &);
+
+
+#endif
@@ -2346,6 +2346,9 @@ extern machine_mode element_mode (const_tree t);
#define DECL_IGNORED_P(NODE) \
(DECL_COMMON_CHECK (NODE)->decl_common.ignored_flag)
+#define DECL_SET_BY_IPA(NODE) \
+ (DECL_COMMON_CHECK (NODE)->decl_common.set_by_ipa)
+
/* Nonzero for a given ..._DECL node means that this node represents an
"abstract instance" of the given declaration (e.g. in the original
declaration of an inline function). When generating symbolic debugging