Skip to content

Commit

Permalink
Fix test for LLVM 20
Browse files Browse the repository at this point in the history
  • Loading branch information
lerno committed Dec 11, 2024
1 parent 723e1dd commit e8ff4af
Showing 1 changed file with 44 additions and 44 deletions.
88 changes: 44 additions & 44 deletions test/test_suite/vector/vector_init_regression.c3t
Original file line number Diff line number Diff line change
Expand Up @@ -120,18 +120,18 @@ entry:
%12 = load float, ptr %x, align 4
%fmul = fmul reassoc arcp contract float %11, %12
%13 = load float, ptr %cosr, align 4
%fpfpext = fpext float %13 to double
%fpfpext = fpext
%fsub = fsub reassoc arcp contract double 1.000000e+00, %fpfpext
%fpfptrunc = fptrunc double %fsub to float
%fpfptrunc = fptrunc
%14 = call reassoc arcp contract float @llvm.fmuladd.f32(float %fmul, float %fpfptrunc, float %10)
%15 = insertelement <4 x float> undef, float %14, i64 0
%16 = load float, ptr %x, align 4
%17 = load float, ptr %y, align 4
%fmul1 = fmul reassoc arcp contract float %16, %17
%18 = load float, ptr %cosr, align 4
%fpfpext2 = fpext float %18 to double
%fpfpext2 = fpext
%fsub3 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext2
%fpfptrunc4 = fptrunc double %fsub3 to float
%fpfptrunc4 = fptrunc
%19 = load float, ptr %z, align 4
%20 = load float, ptr %sinr, align 4
%fmul5 = fmul reassoc arcp contract float %19, %20
Expand All @@ -142,9 +142,9 @@ entry:
%25 = load float, ptr %z, align 4
%fmul6 = fmul reassoc arcp contract float %24, %25
%26 = load float, ptr %cosr, align 4
%fpfpext7 = fpext float %26 to double
%fpfpext7 = fpext
%fsub8 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext7
%fpfptrunc9 = fptrunc double %fsub8 to float
%fpfptrunc9 = fptrunc
%27 = load float, ptr %y, align 4
%28 = load float, ptr %sinr, align 4
%fmul10 = fmul reassoc arcp contract float %27, %28
Expand All @@ -157,9 +157,9 @@ entry:
%33 = load float, ptr %x, align 4
%fmul11 = fmul reassoc arcp contract float %32, %33
%34 = load float, ptr %cosr, align 4
%fpfpext12 = fpext float %34 to double
%fpfpext12 = fpext
%fsub13 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext12
%fpfptrunc14 = fptrunc double %fsub13 to float
%fpfptrunc14 = fptrunc
%35 = load float, ptr %z, align 4
%36 = load float, ptr %sinr, align 4
%fmul15 = fmul reassoc arcp contract float %35, %36
Expand All @@ -170,18 +170,18 @@ entry:
%41 = load float, ptr %y, align 4
%fmul16 = fmul reassoc arcp contract float %40, %41
%42 = load float, ptr %cosr, align 4
%fpfpext17 = fpext float %42 to double
%fpfpext17 = fpext
%fsub18 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext17
%fpfptrunc19 = fptrunc double %fsub18 to float
%fpfptrunc19 = fptrunc
%43 = call reassoc arcp contract float @llvm.fmuladd.f32(float %fmul16, float %fpfptrunc19, float %39)
%44 = insertelement <4 x float> %38, float %43, i64 1
%45 = load float, ptr %y, align 4
%46 = load float, ptr %z, align 4
%fmul20 = fmul reassoc arcp contract float %45, %46
%47 = load float, ptr %cosr, align 4
%fpfpext21 = fpext float %47 to double
%fpfpext21 = fpext
%fsub22 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext21
%fpfptrunc23 = fptrunc double %fsub22 to float
%fpfptrunc23 = fptrunc
%48 = load float, ptr %x, align 4
%49 = load float, ptr %sinr, align 4
%fmul24 = fmul reassoc arcp contract float %48, %49
Expand All @@ -195,9 +195,9 @@ entry:
%55 = load float, ptr %x, align 4
%fmul26 = fmul reassoc arcp contract float %54, %55
%56 = load float, ptr %cosr, align 4
%fpfpext27 = fpext float %56 to double
%fpfpext27 = fpext
%fsub28 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext27
%fpfptrunc29 = fptrunc double %fsub28 to float
%fpfptrunc29 = fptrunc
%57 = load float, ptr %y, align 4
%58 = load float, ptr %sinr, align 4
%fmul30 = fmul reassoc arcp contract float %57, %58
Expand All @@ -208,9 +208,9 @@ entry:
%63 = load float, ptr %y, align 4
%fmul31 = fmul reassoc arcp contract float %62, %63
%64 = load float, ptr %cosr, align 4
%fpfpext32 = fpext float %64 to double
%fpfpext32 = fpext
%fsub33 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext32
%fpfptrunc34 = fptrunc double %fsub33 to float
%fpfptrunc34 = fptrunc
%65 = load float, ptr %x, align 4
%66 = load float, ptr %sinr, align 4
%fmul35 = fmul reassoc arcp contract float %65, %66
Expand All @@ -221,9 +221,9 @@ entry:
%71 = load float, ptr %z, align 4
%fmul36 = fmul reassoc arcp contract float %70, %71
%72 = load float, ptr %cosr, align 4
%fpfpext37 = fpext float %72 to double
%fpfpext37 = fpext
%fsub38 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext37
%fpfptrunc39 = fptrunc double %fsub38 to float
%fpfptrunc39 = fptrunc
%73 = call reassoc arcp contract float @llvm.fmuladd.f32(float %fmul36, float %fpfptrunc39, float %69)
%74 = insertelement <4 x float> %68, float %73, i64 2
%75 = insertelement <4 x float> %74, float 0.000000e+00, i64 3
Expand All @@ -235,18 +235,18 @@ entry:
%78 = load float, ptr %x, align 4
%fmul41 = fmul reassoc arcp contract float %77, %78
%79 = load float, ptr %cosr, align 4
%fpfpext42 = fpext float %79 to double
%fpfpext42 = fpext
%fsub43 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext42
%fpfptrunc44 = fptrunc double %fsub43 to float
%fpfptrunc44 = fptrunc
%80 = call reassoc arcp contract float @llvm.fmuladd.f32(float %fmul41, float %fpfptrunc44, float %76)
%81 = insertelement <4 x float> undef, float %80, i64 0
%82 = load float, ptr %x, align 4
%83 = load float, ptr %y, align 4
%fmul45 = fmul reassoc arcp contract float %82, %83
%84 = load float, ptr %cosr, align 4
%fpfpext46 = fpext float %84 to double
%fpfpext46 = fpext
%fsub47 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext46
%fpfptrunc48 = fptrunc double %fsub47 to float
%fpfptrunc48 = fptrunc
%85 = load float, ptr %z, align 4
%86 = load float, ptr %sinr, align 4
%fmul49 = fmul reassoc arcp contract float %85, %86
Expand All @@ -257,9 +257,9 @@ entry:
%91 = load float, ptr %z, align 4
%fmul50 = fmul reassoc arcp contract float %90, %91
%92 = load float, ptr %cosr, align 4
%fpfpext51 = fpext float %92 to double
%fpfpext51 = fpext
%fsub52 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext51
%fpfptrunc53 = fptrunc double %fsub52 to float
%fpfptrunc53 = fptrunc
%93 = load float, ptr %y, align 4
%94 = load float, ptr %sinr, align 4
%fmul54 = fmul reassoc arcp contract float %93, %94
Expand All @@ -272,9 +272,9 @@ entry:
%99 = load float, ptr %x, align 4
%fmul56 = fmul reassoc arcp contract float %98, %99
%100 = load float, ptr %cosr, align 4
%fpfpext57 = fpext float %100 to double
%fpfpext57 = fpext
%fsub58 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext57
%fpfptrunc59 = fptrunc double %fsub58 to float
%fpfptrunc59 = fptrunc
%101 = load float, ptr %z, align 4
%102 = load float, ptr %sinr, align 4
%fmul60 = fmul reassoc arcp contract float %101, %102
Expand All @@ -285,18 +285,18 @@ entry:
%107 = load float, ptr %y, align 4
%fmul61 = fmul reassoc arcp contract float %106, %107
%108 = load float, ptr %cosr, align 4
%fpfpext62 = fpext float %108 to double
%fpfpext62 = fpext
%fsub63 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext62
%fpfptrunc64 = fptrunc double %fsub63 to float
%fpfptrunc64 = fptrunc
%109 = call reassoc arcp contract float @llvm.fmuladd.f32(float %fmul61, float %fpfptrunc64, float %105)
%110 = insertelement <4 x float> %104, float %109, i64 1
%111 = load float, ptr %y, align 4
%112 = load float, ptr %z, align 4
%fmul65 = fmul reassoc arcp contract float %111, %112
%113 = load float, ptr %cosr, align 4
%fpfpext66 = fpext float %113 to double
%fpfpext66 = fpext
%fsub67 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext66
%fpfptrunc68 = fptrunc double %fsub67 to float
%fpfptrunc68 = fptrunc
%114 = load float, ptr %x, align 4
%115 = load float, ptr %sinr, align 4
%fmul69 = fmul reassoc arcp contract float %114, %115
Expand All @@ -310,9 +310,9 @@ entry:
%121 = load float, ptr %x, align 4
%fmul71 = fmul reassoc arcp contract float %120, %121
%122 = load float, ptr %cosr, align 4
%fpfpext72 = fpext float %122 to double
%fpfpext72 = fpext
%fsub73 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext72
%fpfptrunc74 = fptrunc double %fsub73 to float
%fpfptrunc74 = fptrunc
%123 = load float, ptr %y, align 4
%124 = load float, ptr %sinr, align 4
%fmul75 = fmul reassoc arcp contract float %123, %124
Expand All @@ -323,9 +323,9 @@ entry:
%129 = load float, ptr %y, align 4
%fmul76 = fmul reassoc arcp contract float %128, %129
%130 = load float, ptr %cosr, align 4
%fpfpext77 = fpext float %130 to double
%fpfpext77 = fpext
%fsub78 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext77
%fpfptrunc79 = fptrunc double %fsub78 to float
%fpfptrunc79 = fptrunc
%131 = load float, ptr %x, align 4
%132 = load float, ptr %sinr, align 4
%fmul80 = fmul reassoc arcp contract float %131, %132
Expand All @@ -336,9 +336,9 @@ entry:
%137 = load float, ptr %z, align 4
%fmul81 = fmul reassoc arcp contract float %136, %137
%138 = load float, ptr %cosr, align 4
%fpfpext82 = fpext float %138 to double
%fpfpext82 = fpext
%fsub83 = fsub reassoc arcp contract double 1.000000e+00, %fpfpext82
%fpfptrunc84 = fptrunc double %fsub83 to float
%fpfptrunc84 = fptrunc
%139 = call reassoc arcp contract float @llvm.fmuladd.f32(float %fmul81, float %fpfptrunc84, float %135)
%140 = insertelement <4 x float> %134, float %139, i64 2
%141 = insertelement <4 x float> %140, float 0.000000e+00, i64 3
Expand All @@ -360,16 +360,16 @@ loop.body: ; preds = %loop.cond
store <4 x float> %144, ptr %v, align 16
%145 = load <4 x float>, ptr %v, align 16
%146 = extractelement <4 x float> %145, i64 0
%fpfpext86 = fpext float %146 to double
%fpfpext86 = fpext
%147 = load <4 x float>, ptr %v, align 16
%148 = extractelement <4 x float> %147, i64 1
%fpfpext87 = fpext float %148 to double
%fpfpext87 = fpext
%149 = load <4 x float>, ptr %v, align 16
%150 = extractelement <4 x float> %149, i64 2
%fpfpext88 = fpext float %150 to double
%fpfpext88 = fpext
%151 = load <4 x float>, ptr %v, align 16
%152 = extractelement <4 x float> %151, i64 3
%fpfpext89 = fpext float %152 to double
%fpfpext89 = fpext
%153 = call i32 (ptr, ...) @printf(ptr @.str, double %fpfpext86, double %fpfpext87, double %fpfpext88, double %fpfpext89)
%154 = load i64, ptr %.anon, align 8
%addnuw = add nuw i64 %154, 1
Expand All @@ -393,16 +393,16 @@ loop.body93: ; preds = %loop.cond91
store <4 x float> %158, ptr %v94, align 16
%159 = load <4 x float>, ptr %v94, align 16
%160 = extractelement <4 x float> %159, i64 0
%fpfpext96 = fpext float %160 to double
%fpfpext96 = fpext
%161 = load <4 x float>, ptr %v94, align 16
%162 = extractelement <4 x float> %161, i64 1
%fpfpext97 = fpext float %162 to double
%fpfpext97 = fpext
%163 = load <4 x float>, ptr %v94, align 16
%164 = extractelement <4 x float> %163, i64 2
%fpfpext98 = fpext float %164 to double
%fpfpext98 = fpext
%165 = load <4 x float>, ptr %v94, align 16
%166 = extractelement <4 x float> %165, i64 3
%fpfpext99 = fpext float %166 to double
%fpfpext99 = fpext
%167 = call i32 (ptr, ...) @printf(ptr @.str.2, double %fpfpext96, double %fpfpext97, double %fpfpext98, double %fpfpext99)
%168 = load i64, ptr %.anon90, align 8
%addnuw100 = add nuw i64 %168, 1
Expand Down

0 comments on commit e8ff4af

Please sign in to comment.