@@ -5,8 +5,8 @@ double __GIMPLE (ssa, startwith("loop"))
neg_xi (double *x)
{
int i;
- long unsigned int index;
- long unsigned int offset;
+ __SIZETYPE__ index;
+ __SIZETYPE__ offset;
double * xi_ptr;
double xi;
double neg_xi;
@@ -20,8 +20,8 @@ neg_xi (double *x)
res_1 = __PHI (__BB5: 0.0, __BB3: res_2);
i_4 = __PHI (__BB5: 0, __BB3: i_5);
ivtmp_6 = __PHI (__BB5: 100U, __BB3: ivtmp_7);
- index = (long unsigned int) i_4;
- offset = index * 8UL;
+ index = (__SIZETYPE__ ) i_4;
+ offset = index * _Literal (__SIZETYPE__) 8;
xi_ptr = x_8(D) + offset;
xi = *xi_ptr;
neg_xi = -xi;
Even though the aarch64-mingw32 support has not been committed yet, we should fix some of the testcases. In this case gcc.target/aarch64/sve/loop_add_6.c is easy to fix. We should use __SIZETYPE__ instead of `unsigned long` for the variables that will be used for pointer plus. Committed as obvious after a quick test on aarch64-linux-gnu. gcc/testsuite/ChangeLog: PR testsuite/114177 * gcc.target/aarch64/sve/loop_add_6.c: Use __SIZETYPE__ instead of `unsigned long` for index and offset variables. Signed-off-by: Andrew Pinski <quic_apinski@quicinc.com> --- gcc/testsuite/gcc.target/aarch64/sve/loop_add_6.c | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-)