From e8ff4af5b95f0d0e2d5e0fd047618684cb66fce9 Mon Sep 17 00:00:00 2001 From: Christoffer Lerno Date: Wed, 11 Dec 2024 23:05:57 +0100 Subject: [PATCH] Fix test for LLVM 20 --- .../vector/vector_init_regression.c3t | 88 +++++++++---------- 1 file changed, 44 insertions(+), 44 deletions(-) diff --git a/test/test_suite/vector/vector_init_regression.c3t b/test/test_suite/vector/vector_init_regression.c3t index d6120d4e8..73ad3b1e0 100644 --- a/test/test_suite/vector/vector_init_regression.c3t +++ b/test/test_suite/vector/vector_init_regression.c3t @@ -120,18 +120,18 @@ entry: %12 = load float, ptr %x, align 4 %fmul = fmul reassoc arcp contract float %11, %12 %13 = load float, ptr %cosr, align 4 - %fpfpext = fpext float %13 to double + %fpfpext = fpext %fsub = fsub reassoc arcp contract double 1.000000e+00, %fpfpext - %fpfptrunc = fptrunc double %fsub to float + %fpfptrunc = fptrunc %14 = call reassoc arcp contract float @llvm.fmuladd.f32(float %fmul, float %fpfptrunc, float %10) %15 = insertelement <4 x float> undef, float %14, i64 0 %16 = load float, ptr %x, align 4 %17 = load float, ptr %y, align 4 %fmul1 = fmul reassoc arcp contract float %16, %17 %18 = load float, ptr %cosr, align 4 - %fpfpext2 = fpext float %18 to double + %fpfpext2 = fpext %fsub3 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext2 - %fpfptrunc4 = fptrunc double %fsub3 to float + %fpfptrunc4 = fptrunc %19 = load float, ptr %z, align 4 %20 = load float, ptr %sinr, align 4 %fmul5 = fmul reassoc arcp contract float %19, %20 @@ -142,9 +142,9 @@ entry: %25 = load float, ptr %z, align 4 %fmul6 = fmul reassoc arcp contract float %24, %25 %26 = load float, ptr %cosr, align 4 - %fpfpext7 = fpext float %26 to double + %fpfpext7 = fpext %fsub8 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext7 - %fpfptrunc9 = fptrunc double %fsub8 to float + %fpfptrunc9 = fptrunc %27 = load float, ptr %y, align 4 %28 = load float, ptr %sinr, align 4 %fmul10 = fmul reassoc arcp contract float %27, %28 @@ -157,9 +157,9 @@ entry: %33 = load float, ptr %x, align 4 %fmul11 = fmul reassoc arcp contract float %32, %33 %34 = load float, ptr %cosr, align 4 - %fpfpext12 = fpext float %34 to double + %fpfpext12 = fpext %fsub13 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext12 - %fpfptrunc14 = fptrunc double %fsub13 to float + %fpfptrunc14 = fptrunc %35 = load float, ptr %z, align 4 %36 = load float, ptr %sinr, align 4 %fmul15 = fmul reassoc arcp contract float %35, %36 @@ -170,18 +170,18 @@ entry: %41 = load float, ptr %y, align 4 %fmul16 = fmul reassoc arcp contract float %40, %41 %42 = load float, ptr %cosr, align 4 - %fpfpext17 = fpext float %42 to double + %fpfpext17 = fpext %fsub18 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext17 - %fpfptrunc19 = fptrunc double %fsub18 to float + %fpfptrunc19 = fptrunc %43 = call reassoc arcp contract float @llvm.fmuladd.f32(float %fmul16, float %fpfptrunc19, float %39) %44 = insertelement <4 x float> %38, float %43, i64 1 %45 = load float, ptr %y, align 4 %46 = load float, ptr %z, align 4 %fmul20 = fmul reassoc arcp contract float %45, %46 %47 = load float, ptr %cosr, align 4 - %fpfpext21 = fpext float %47 to double + %fpfpext21 = fpext %fsub22 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext21 - %fpfptrunc23 = fptrunc double %fsub22 to float + %fpfptrunc23 = fptrunc %48 = load float, ptr %x, align 4 %49 = load float, ptr %sinr, align 4 %fmul24 = fmul reassoc arcp contract float %48, %49 @@ -195,9 +195,9 @@ entry: %55 = load float, ptr %x, align 4 %fmul26 = fmul reassoc arcp contract float %54, %55 %56 = load float, ptr %cosr, align 4 - %fpfpext27 = fpext float %56 to double + %fpfpext27 = fpext %fsub28 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext27 - %fpfptrunc29 = fptrunc double %fsub28 to float + %fpfptrunc29 = fptrunc %57 = load float, ptr %y, align 4 %58 = load float, ptr %sinr, align 4 %fmul30 = fmul reassoc arcp contract float %57, %58 @@ -208,9 +208,9 @@ entry: %63 = load float, ptr %y, align 4 %fmul31 = fmul reassoc arcp contract float %62, %63 %64 = load float, ptr %cosr, align 4 - %fpfpext32 = fpext float %64 to double + %fpfpext32 = fpext %fsub33 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext32 - %fpfptrunc34 = fptrunc double %fsub33 to float + %fpfptrunc34 = fptrunc %65 = load float, ptr %x, align 4 %66 = load float, ptr %sinr, align 4 %fmul35 = fmul reassoc arcp contract float %65, %66 @@ -221,9 +221,9 @@ entry: %71 = load float, ptr %z, align 4 %fmul36 = fmul reassoc arcp contract float %70, %71 %72 = load float, ptr %cosr, align 4 - %fpfpext37 = fpext float %72 to double + %fpfpext37 = fpext %fsub38 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext37 - %fpfptrunc39 = fptrunc double %fsub38 to float + %fpfptrunc39 = fptrunc %73 = call reassoc arcp contract float @llvm.fmuladd.f32(float %fmul36, float %fpfptrunc39, float %69) %74 = insertelement <4 x float> %68, float %73, i64 2 %75 = insertelement <4 x float> %74, float 0.000000e+00, i64 3 @@ -235,18 +235,18 @@ entry: %78 = load float, ptr %x, align 4 %fmul41 = fmul reassoc arcp contract float %77, %78 %79 = load float, ptr %cosr, align 4 - %fpfpext42 = fpext float %79 to double + %fpfpext42 = fpext %fsub43 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext42 - %fpfptrunc44 = fptrunc double %fsub43 to float + %fpfptrunc44 = fptrunc %80 = call reassoc arcp contract float @llvm.fmuladd.f32(float %fmul41, float %fpfptrunc44, float %76) %81 = insertelement <4 x float> undef, float %80, i64 0 %82 = load float, ptr %x, align 4 %83 = load float, ptr %y, align 4 %fmul45 = fmul reassoc arcp contract float %82, %83 %84 = load float, ptr %cosr, align 4 - %fpfpext46 = fpext float %84 to double + %fpfpext46 = fpext %fsub47 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext46 - %fpfptrunc48 = fptrunc double %fsub47 to float + %fpfptrunc48 = fptrunc %85 = load float, ptr %z, align 4 %86 = load float, ptr %sinr, align 4 %fmul49 = fmul reassoc arcp contract float %85, %86 @@ -257,9 +257,9 @@ entry: %91 = load float, ptr %z, align 4 %fmul50 = fmul reassoc arcp contract float %90, %91 %92 = load float, ptr %cosr, align 4 - %fpfpext51 = fpext float %92 to double + %fpfpext51 = fpext %fsub52 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext51 - %fpfptrunc53 = fptrunc double %fsub52 to float + %fpfptrunc53 = fptrunc %93 = load float, ptr %y, align 4 %94 = load float, ptr %sinr, align 4 %fmul54 = fmul reassoc arcp contract float %93, %94 @@ -272,9 +272,9 @@ entry: %99 = load float, ptr %x, align 4 %fmul56 = fmul reassoc arcp contract float %98, %99 %100 = load float, ptr %cosr, align 4 - %fpfpext57 = fpext float %100 to double + %fpfpext57 = fpext %fsub58 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext57 - %fpfptrunc59 = fptrunc double %fsub58 to float + %fpfptrunc59 = fptrunc %101 = load float, ptr %z, align 4 %102 = load float, ptr %sinr, align 4 %fmul60 = fmul reassoc arcp contract float %101, %102 @@ -285,18 +285,18 @@ entry: %107 = load float, ptr %y, align 4 %fmul61 = fmul reassoc arcp contract float %106, %107 %108 = load float, ptr %cosr, align 4 - %fpfpext62 = fpext float %108 to double + %fpfpext62 = fpext %fsub63 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext62 - %fpfptrunc64 = fptrunc double %fsub63 to float + %fpfptrunc64 = fptrunc %109 = call reassoc arcp contract float @llvm.fmuladd.f32(float %fmul61, float %fpfptrunc64, float %105) %110 = insertelement <4 x float> %104, float %109, i64 1 %111 = load float, ptr %y, align 4 %112 = load float, ptr %z, align 4 %fmul65 = fmul reassoc arcp contract float %111, %112 %113 = load float, ptr %cosr, align 4 - %fpfpext66 = fpext float %113 to double + %fpfpext66 = fpext %fsub67 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext66 - %fpfptrunc68 = fptrunc double %fsub67 to float + %fpfptrunc68 = fptrunc %114 = load float, ptr %x, align 4 %115 = load float, ptr %sinr, align 4 %fmul69 = fmul reassoc arcp contract float %114, %115 @@ -310,9 +310,9 @@ entry: %121 = load float, ptr %x, align 4 %fmul71 = fmul reassoc arcp contract float %120, %121 %122 = load float, ptr %cosr, align 4 - %fpfpext72 = fpext float %122 to double + %fpfpext72 = fpext %fsub73 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext72 - %fpfptrunc74 = fptrunc double %fsub73 to float + %fpfptrunc74 = fptrunc %123 = load float, ptr %y, align 4 %124 = load float, ptr %sinr, align 4 %fmul75 = fmul reassoc arcp contract float %123, %124 @@ -323,9 +323,9 @@ entry: %129 = load float, ptr %y, align 4 %fmul76 = fmul reassoc arcp contract float %128, %129 %130 = load float, ptr %cosr, align 4 - %fpfpext77 = fpext float %130 to double + %fpfpext77 = fpext %fsub78 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext77 - %fpfptrunc79 = fptrunc double %fsub78 to float + %fpfptrunc79 = fptrunc %131 = load float, ptr %x, align 4 %132 = load float, ptr %sinr, align 4 %fmul80 = fmul reassoc arcp contract float %131, %132 @@ -336,9 +336,9 @@ entry: %137 = load float, ptr %z, align 4 %fmul81 = fmul reassoc arcp contract float %136, %137 %138 = load float, ptr %cosr, align 4 - %fpfpext82 = fpext float %138 to double + %fpfpext82 = fpext %fsub83 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext82 - %fpfptrunc84 = fptrunc double %fsub83 to float + %fpfptrunc84 = fptrunc %139 = call reassoc arcp contract float @llvm.fmuladd.f32(float %fmul81, float %fpfptrunc84, float %135) %140 = insertelement <4 x float> %134, float %139, i64 2 %141 = insertelement <4 x float> %140, float 0.000000e+00, i64 3 @@ -360,16 +360,16 @@ loop.body: ; preds = %loop.cond store <4 x float> %144, ptr %v, align 16 %145 = load <4 x float>, ptr %v, align 16 %146 = extractelement <4 x float> %145, i64 0 - %fpfpext86 = fpext float %146 to double + %fpfpext86 = fpext %147 = load <4 x float>, ptr %v, align 16 %148 = extractelement <4 x float> %147, i64 1 - %fpfpext87 = fpext float %148 to double + %fpfpext87 = fpext %149 = load <4 x float>, ptr %v, align 16 %150 = extractelement <4 x float> %149, i64 2 - %fpfpext88 = fpext float %150 to double + %fpfpext88 = fpext %151 = load <4 x float>, ptr %v, align 16 %152 = extractelement <4 x float> %151, i64 3 - %fpfpext89 = fpext float %152 to double + %fpfpext89 = fpext %153 = call i32 (ptr, ...) @printf(ptr @.str, double %fpfpext86, double %fpfpext87, double %fpfpext88, double %fpfpext89) %154 = load i64, ptr %.anon, align 8 %addnuw = add nuw i64 %154, 1 @@ -393,16 +393,16 @@ loop.body93: ; preds = %loop.cond91 store <4 x float> %158, ptr %v94, align 16 %159 = load <4 x float>, ptr %v94, align 16 %160 = extractelement <4 x float> %159, i64 0 - %fpfpext96 = fpext float %160 to double + %fpfpext96 = fpext %161 = load <4 x float>, ptr %v94, align 16 %162 = extractelement <4 x float> %161, i64 1 - %fpfpext97 = fpext float %162 to double + %fpfpext97 = fpext %163 = load <4 x float>, ptr %v94, align 16 %164 = extractelement <4 x float> %163, i64 2 - %fpfpext98 = fpext float %164 to double + %fpfpext98 = fpext %165 = load <4 x float>, ptr %v94, align 16 %166 = extractelement <4 x float> %165, i64 3 - %fpfpext99 = fpext float %166 to double + %fpfpext99 = fpext %167 = call i32 (ptr, ...) @printf(ptr @.str.2, double %fpfpext96, double %fpfpext97, double %fpfpext98, double %fpfpext99) %168 = load i64, ptr %.anon90, align 8 %addnuw100 = add nuw i64 %168, 1