@@ -1138,8 +1138,7 @@ conversion_warning (location_t loc, tree
conversion_kind = unsafe_conversion_p (loc, type, expr, result, true);
if (conversion_kind == UNSAFE_IMAGINARY)
warning_at (loc, OPT_Wconversion,
- "conversion from %qT to to %qT discards imaginary "
- "component",
+ "conversion from %qT to %qT discards imaginary component",
expr_type, type);
else
{
@@ -691,7 +691,7 @@ mem_type_for_type (BrigType16_t type)
/* HSA has non-intuitive constraints on load/store types. If it's
a bit-type it _must_ be B128, if it's not a bit-type it must be
64bit max. So for loading entities of 128 bits (e.g. vectors)
- we have to to B128, while for loading the rest we have to use the
+ we have to use B128, while for loading the rest we have to use the
input type (??? or maybe also flattened to a equally sized non-vector
unsigned type?). */
if ((type & BRIG_TYPE_PACK_MASK) == BRIG_TYPE_PACK_128)
@@ -2700,7 +2700,7 @@ vect_do_peeling (loop_vec_info loop_vinf
/* Function vect_create_cond_for_niters_checks.
Create a conditional expression that represents the run-time checks for
- loop's niter. The loop is guaranteed to to terminate if the run-time
+ loop's niter. The loop is guaranteed to terminate if the run-time
checks hold.
Input:
@@ -417,7 +417,7 @@ builtin_memref::set_base_and_offset (tre
poly_int64 bytepos = exact_div (bitpos, BITS_PER_UNIT);
- /* Convert the poly_int64 offset to to offset_int. The offset
+ /* Convert the poly_int64 offset to offset_int. The offset
should be constant but be prepared for it not to be just in
case. */
offset_int cstoff;