xref: /llvm-project/llvm/test/CodeGen/RISCV/pr69586.ll (revision 9122c5235ec85ce0c0ad337e862b006e7b349d84)
1; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 3
2; RUN: llc < %s -mtriple=riscv64 -mattr=+m,+v,+xsfvcp \
3; RUN:   -riscv-use-rematerializable-movimm=false | FileCheck %s --check-prefix=NOREMAT
4; RUN: llc < %s -mtriple=riscv64 -mattr=+m,+v,+xsfvcp \
5; RUN:   --riscv-use-rematerializable-movimm=true | FileCheck %s --check-prefix=REMAT
6
7define void @test(ptr %0, ptr %1, i64 %2) {
8; NOREMAT-LABEL: test:
9; NOREMAT:       # %bb.0:
10; NOREMAT-NEXT:    addi sp, sp, -752
11; NOREMAT-NEXT:    .cfi_def_cfa_offset 752
12; NOREMAT-NEXT:    sd ra, 744(sp) # 8-byte Folded Spill
13; NOREMAT-NEXT:    sd s0, 736(sp) # 8-byte Folded Spill
14; NOREMAT-NEXT:    sd s1, 728(sp) # 8-byte Folded Spill
15; NOREMAT-NEXT:    sd s2, 720(sp) # 8-byte Folded Spill
16; NOREMAT-NEXT:    sd s3, 712(sp) # 8-byte Folded Spill
17; NOREMAT-NEXT:    sd s4, 704(sp) # 8-byte Folded Spill
18; NOREMAT-NEXT:    sd s5, 696(sp) # 8-byte Folded Spill
19; NOREMAT-NEXT:    sd s6, 688(sp) # 8-byte Folded Spill
20; NOREMAT-NEXT:    sd s7, 680(sp) # 8-byte Folded Spill
21; NOREMAT-NEXT:    sd s8, 672(sp) # 8-byte Folded Spill
22; NOREMAT-NEXT:    sd s9, 664(sp) # 8-byte Folded Spill
23; NOREMAT-NEXT:    sd s10, 656(sp) # 8-byte Folded Spill
24; NOREMAT-NEXT:    sd s11, 648(sp) # 8-byte Folded Spill
25; NOREMAT-NEXT:    .cfi_offset ra, -8
26; NOREMAT-NEXT:    .cfi_offset s0, -16
27; NOREMAT-NEXT:    .cfi_offset s1, -24
28; NOREMAT-NEXT:    .cfi_offset s2, -32
29; NOREMAT-NEXT:    .cfi_offset s3, -40
30; NOREMAT-NEXT:    .cfi_offset s4, -48
31; NOREMAT-NEXT:    .cfi_offset s5, -56
32; NOREMAT-NEXT:    .cfi_offset s6, -64
33; NOREMAT-NEXT:    .cfi_offset s7, -72
34; NOREMAT-NEXT:    .cfi_offset s8, -80
35; NOREMAT-NEXT:    .cfi_offset s9, -88
36; NOREMAT-NEXT:    .cfi_offset s10, -96
37; NOREMAT-NEXT:    .cfi_offset s11, -104
38; NOREMAT-NEXT:    csrr a2, vlenb
39; NOREMAT-NEXT:    slli a2, a2, 1
40; NOREMAT-NEXT:    sub sp, sp, a2
41; NOREMAT-NEXT:    .cfi_escape 0x0f, 0x0e, 0x72, 0x00, 0x11, 0xf0, 0x05, 0x22, 0x11, 0x02, 0x92, 0xa2, 0x38, 0x00, 0x1e, 0x22 # sp + 752 + 2 * vlenb
42; NOREMAT-NEXT:    mv a7, a0
43; NOREMAT-NEXT:    li a0, 32
44; NOREMAT-NEXT:    addi a5, a7, 512
45; NOREMAT-NEXT:    addi a4, a7, 1024
46; NOREMAT-NEXT:    addi a6, a7, 1536
47; NOREMAT-NEXT:    li t4, 1
48; NOREMAT-NEXT:    li a2, 5
49; NOREMAT-NEXT:    li t1, 3
50; NOREMAT-NEXT:    li t0, 7
51; NOREMAT-NEXT:    lui t5, 1
52; NOREMAT-NEXT:    li s4, 9
53; NOREMAT-NEXT:    li s6, 11
54; NOREMAT-NEXT:    li s9, 13
55; NOREMAT-NEXT:    li ra, 15
56; NOREMAT-NEXT:    lui t2, 2
57; NOREMAT-NEXT:    lui s1, 3
58; NOREMAT-NEXT:    lui t3, 4
59; NOREMAT-NEXT:    lui s0, 5
60; NOREMAT-NEXT:    lui s3, 6
61; NOREMAT-NEXT:    lui s7, 7
62; NOREMAT-NEXT:    vsetvli zero, a0, e32, m2, ta, ma
63; NOREMAT-NEXT:    slli t4, t4, 11
64; NOREMAT-NEXT:    sd t4, 512(sp) # 8-byte Folded Spill
65; NOREMAT-NEXT:    slli a3, a2, 9
66; NOREMAT-NEXT:    sd a3, 504(sp) # 8-byte Folded Spill
67; NOREMAT-NEXT:    slli t6, t1, 10
68; NOREMAT-NEXT:    slli s2, t0, 9
69; NOREMAT-NEXT:    add a0, a7, t5
70; NOREMAT-NEXT:    lui s11, 1
71; NOREMAT-NEXT:    slli s4, s4, 9
72; NOREMAT-NEXT:    slli s5, a2, 10
73; NOREMAT-NEXT:    slli s6, s6, 9
74; NOREMAT-NEXT:    slli s8, t1, 11
75; NOREMAT-NEXT:    vle32.v v8, (a5)
76; NOREMAT-NEXT:    slli s9, s9, 9
77; NOREMAT-NEXT:    li t5, 13
78; NOREMAT-NEXT:    vle32.v v10, (a4)
79; NOREMAT-NEXT:    vle32.v v2, (a4)
80; NOREMAT-NEXT:    slli s10, t0, 10
81; NOREMAT-NEXT:    vle32.v v0, (a6)
82; NOREMAT-NEXT:    vle32.v v12, (a6)
83; NOREMAT-NEXT:    slli ra, ra, 9
84; NOREMAT-NEXT:    vle32.v v4, (a0)
85; NOREMAT-NEXT:    vle32.v v20, (a0)
86; NOREMAT-NEXT:    add a4, a7, t2
87; NOREMAT-NEXT:    vle32.v v6, (a4)
88; NOREMAT-NEXT:    vle32.v v30, (a4)
89; NOREMAT-NEXT:    add a4, a7, s1
90; NOREMAT-NEXT:    vle32.v v28, (a4)
91; NOREMAT-NEXT:    vle32.v v26, (a4)
92; NOREMAT-NEXT:    add a4, a7, t3
93; NOREMAT-NEXT:    vle32.v v24, (a4)
94; NOREMAT-NEXT:    vle32.v v22, (a4)
95; NOREMAT-NEXT:    add a4, a7, s0
96; NOREMAT-NEXT:    vle32.v v14, (a7)
97; NOREMAT-NEXT:    vle32.v v18, (a4)
98; NOREMAT-NEXT:    vle32.v v16, (a4)
99; NOREMAT-NEXT:    add a4, a7, s3
100; NOREMAT-NEXT:    sf.vc.vv 3, 0, v14, v8
101; NOREMAT-NEXT:    vle32.v v14, (a4)
102; NOREMAT-NEXT:    sf.vc.vv 3, 0, v8, v10
103; NOREMAT-NEXT:    vle32.v v8, (a4)
104; NOREMAT-NEXT:    addi a0, sp, 640
105; NOREMAT-NEXT:    vs2r.v v8, (a0) # Unknown-size Folded Spill
106; NOREMAT-NEXT:    add a4, a7, t4
107; NOREMAT-NEXT:    vle32.v v10, (a4)
108; NOREMAT-NEXT:    sf.vc.vv 3, 0, v2, v0
109; NOREMAT-NEXT:    vle32.v v2, (a4)
110; NOREMAT-NEXT:    add a4, a7, a3
111; NOREMAT-NEXT:    vle32.v v0, (a4)
112; NOREMAT-NEXT:    sf.vc.vv 3, 0, v12, v10
113; NOREMAT-NEXT:    vle32.v v10, (a4)
114; NOREMAT-NEXT:    add a4, a7, t6
115; NOREMAT-NEXT:    vle32.v v12, (a4)
116; NOREMAT-NEXT:    sf.vc.vv 3, 0, v2, v0
117; NOREMAT-NEXT:    vle32.v v2, (a4)
118; NOREMAT-NEXT:    add a4, a7, s2
119; NOREMAT-NEXT:    vle32.v v8, (a4)
120; NOREMAT-NEXT:    sf.vc.vv 3, 0, v10, v12
121; NOREMAT-NEXT:    vle32.v v12, (a4)
122; NOREMAT-NEXT:    add a4, a7, s7
123; NOREMAT-NEXT:    vle32.v v0, (a4)
124; NOREMAT-NEXT:    sf.vc.vv 3, 0, v2, v8
125; NOREMAT-NEXT:    vle32.v v10, (a4)
126; NOREMAT-NEXT:    add a4, a7, s4
127; NOREMAT-NEXT:    vle32.v v8, (a4)
128; NOREMAT-NEXT:    sf.vc.vv 3, 0, v12, v4
129; NOREMAT-NEXT:    vle32.v v12, (a4)
130; NOREMAT-NEXT:    add a4, a7, s5
131; NOREMAT-NEXT:    vle32.v v4, (a4)
132; NOREMAT-NEXT:    sf.vc.vv 3, 0, v20, v8
133; NOREMAT-NEXT:    vle32.v v8, (a4)
134; NOREMAT-NEXT:    add a4, a7, s6
135; NOREMAT-NEXT:    vle32.v v20, (a4)
136; NOREMAT-NEXT:    sf.vc.vv 3, 0, v12, v4
137; NOREMAT-NEXT:    vle32.v v12, (a4)
138; NOREMAT-NEXT:    add a4, a7, s8
139; NOREMAT-NEXT:    vle32.v v4, (a4)
140; NOREMAT-NEXT:    sf.vc.vv 3, 0, v8, v20
141; NOREMAT-NEXT:    vle32.v v8, (a4)
142; NOREMAT-NEXT:    add a4, a7, s9
143; NOREMAT-NEXT:    vle32.v v20, (a4)
144; NOREMAT-NEXT:    sf.vc.vv 3, 0, v12, v4
145; NOREMAT-NEXT:    vle32.v v12, (a4)
146; NOREMAT-NEXT:    add a4, a7, s10
147; NOREMAT-NEXT:    vle32.v v4, (a4)
148; NOREMAT-NEXT:    sf.vc.vv 3, 0, v8, v20
149; NOREMAT-NEXT:    vle32.v v8, (a4)
150; NOREMAT-NEXT:    add a4, a7, ra
151; NOREMAT-NEXT:    vle32.v v2, (a4)
152; NOREMAT-NEXT:    sf.vc.vv 3, 0, v12, v4
153; NOREMAT-NEXT:    lui t4, 8
154; NOREMAT-NEXT:    add a5, a7, t4
155; NOREMAT-NEXT:    vle32.v v20, (a5)
156; NOREMAT-NEXT:    vle32.v v12, (a4)
157; NOREMAT-NEXT:    sf.vc.vv 3, 0, v8, v2
158; NOREMAT-NEXT:    li a4, 17
159; NOREMAT-NEXT:    slli a4, a4, 9
160; NOREMAT-NEXT:    li s1, 17
161; NOREMAT-NEXT:    sd a4, 624(sp) # 8-byte Folded Spill
162; NOREMAT-NEXT:    add a4, a7, a4
163; NOREMAT-NEXT:    vle32.v v8, (a4)
164; NOREMAT-NEXT:    vle32.v v4, (a4)
165; NOREMAT-NEXT:    sf.vc.vv 3, 0, v12, v6
166; NOREMAT-NEXT:    li a5, 9
167; NOREMAT-NEXT:    slli a4, a5, 10
168; NOREMAT-NEXT:    sd a4, 616(sp) # 8-byte Folded Spill
169; NOREMAT-NEXT:    add a4, a7, a4
170; NOREMAT-NEXT:    vle32.v v12, (a4)
171; NOREMAT-NEXT:    vle32.v v6, (a4)
172; NOREMAT-NEXT:    sf.vc.vv 3, 0, v30, v8
173; NOREMAT-NEXT:    li a4, 19
174; NOREMAT-NEXT:    slli a4, a4, 9
175; NOREMAT-NEXT:    li t2, 19
176; NOREMAT-NEXT:    sd a4, 608(sp) # 8-byte Folded Spill
177; NOREMAT-NEXT:    add a4, a7, a4
178; NOREMAT-NEXT:    vle32.v v8, (a4)
179; NOREMAT-NEXT:    vle32.v v30, (a4)
180; NOREMAT-NEXT:    slli a3, a2, 11
181; NOREMAT-NEXT:    sd a3, 600(sp) # 8-byte Folded Spill
182; NOREMAT-NEXT:    sf.vc.vv 3, 0, v4, v12
183; NOREMAT-NEXT:    add a3, a7, a3
184; NOREMAT-NEXT:    vle32.v v12, (a3)
185; NOREMAT-NEXT:    vle32.v v4, (a3)
186; NOREMAT-NEXT:    sf.vc.vv 3, 0, v6, v8
187; NOREMAT-NEXT:    li s7, 21
188; NOREMAT-NEXT:    slli a3, s7, 9
189; NOREMAT-NEXT:    sd a3, 592(sp) # 8-byte Folded Spill
190; NOREMAT-NEXT:    add a3, a7, a3
191; NOREMAT-NEXT:    vle32.v v8, (a3)
192; NOREMAT-NEXT:    vle32.v v6, (a3)
193; NOREMAT-NEXT:    sf.vc.vv 3, 0, v30, v12
194; NOREMAT-NEXT:    li a6, 11
195; NOREMAT-NEXT:    slli a3, a6, 10
196; NOREMAT-NEXT:    sd a3, 584(sp) # 8-byte Folded Spill
197; NOREMAT-NEXT:    add a3, a7, a3
198; NOREMAT-NEXT:    vle32.v v12, (a3)
199; NOREMAT-NEXT:    vle32.v v30, (a3)
200; NOREMAT-NEXT:    sf.vc.vv 3, 0, v4, v8
201; NOREMAT-NEXT:    li s3, 23
202; NOREMAT-NEXT:    slli a3, s3, 9
203; NOREMAT-NEXT:    sd a3, 576(sp) # 8-byte Folded Spill
204; NOREMAT-NEXT:    add a3, a7, a3
205; NOREMAT-NEXT:    vle32.v v8, (a3)
206; NOREMAT-NEXT:    vle32.v v4, (a3)
207; NOREMAT-NEXT:    sf.vc.vv 3, 0, v6, v12
208; NOREMAT-NEXT:    li s0, 25
209; NOREMAT-NEXT:    slli a3, s0, 9
210; NOREMAT-NEXT:    sd a3, 568(sp) # 8-byte Folded Spill
211; NOREMAT-NEXT:    add a3, a7, a3
212; NOREMAT-NEXT:    vle32.v v12, (a3)
213; NOREMAT-NEXT:    vle32.v v6, (a3)
214; NOREMAT-NEXT:    sf.vc.vv 3, 0, v30, v8
215; NOREMAT-NEXT:    slli a3, t5, 10
216; NOREMAT-NEXT:    sd a3, 560(sp) # 8-byte Folded Spill
217; NOREMAT-NEXT:    add a3, a7, a3
218; NOREMAT-NEXT:    vle32.v v8, (a3)
219; NOREMAT-NEXT:    vle32.v v30, (a3)
220; NOREMAT-NEXT:    sf.vc.vv 3, 0, v4, v28
221; NOREMAT-NEXT:    li t3, 27
222; NOREMAT-NEXT:    slli a3, t3, 9
223; NOREMAT-NEXT:    sd a3, 552(sp) # 8-byte Folded Spill
224; NOREMAT-NEXT:    add a3, a7, a3
225; NOREMAT-NEXT:    vle32.v v28, (a3)
226; NOREMAT-NEXT:    vle32.v v4, (a3)
227; NOREMAT-NEXT:    slli a2, t0, 11
228; NOREMAT-NEXT:    sd a2, 544(sp) # 8-byte Folded Spill
229; NOREMAT-NEXT:    sf.vc.vv 3, 0, v26, v12
230; NOREMAT-NEXT:    add a2, a7, a2
231; NOREMAT-NEXT:    vle32.v v12, (a2)
232; NOREMAT-NEXT:    vle32.v v26, (a2)
233; NOREMAT-NEXT:    sf.vc.vv 3, 0, v6, v8
234; NOREMAT-NEXT:    li t0, 29
235; NOREMAT-NEXT:    slli a2, t0, 9
236; NOREMAT-NEXT:    sd a2, 536(sp) # 8-byte Folded Spill
237; NOREMAT-NEXT:    add a2, a7, a2
238; NOREMAT-NEXT:    vle32.v v8, (a2)
239; NOREMAT-NEXT:    vle32.v v6, (a2)
240; NOREMAT-NEXT:    sf.vc.vv 3, 0, v30, v28
241; NOREMAT-NEXT:    li a3, 15
242; NOREMAT-NEXT:    slli a2, a3, 10
243; NOREMAT-NEXT:    sd a2, 528(sp) # 8-byte Folded Spill
244; NOREMAT-NEXT:    add a2, a7, a2
245; NOREMAT-NEXT:    vle32.v v28, (a2)
246; NOREMAT-NEXT:    vle32.v v30, (a2)
247; NOREMAT-NEXT:    sf.vc.vv 3, 0, v4, v12
248; NOREMAT-NEXT:    li t1, 31
249; NOREMAT-NEXT:    slli a2, t1, 9
250; NOREMAT-NEXT:    sd a2, 520(sp) # 8-byte Folded Spill
251; NOREMAT-NEXT:    add a2, a7, a2
252; NOREMAT-NEXT:    vle32.v v12, (a2)
253; NOREMAT-NEXT:    vle32.v v4, (a2)
254; NOREMAT-NEXT:    sf.vc.vv 3, 0, v26, v8
255; NOREMAT-NEXT:    lui a4, 4
256; NOREMAT-NEXT:    addiw a0, a4, 512
257; NOREMAT-NEXT:    sd a0, 496(sp) # 8-byte Folded Spill
258; NOREMAT-NEXT:    add a0, a7, a0
259; NOREMAT-NEXT:    vle32.v v8, (a0)
260; NOREMAT-NEXT:    vle32.v v26, (a0)
261; NOREMAT-NEXT:    sf.vc.vv 3, 0, v6, v28
262; NOREMAT-NEXT:    slli a2, s1, 10
263; NOREMAT-NEXT:    sd a2, 488(sp) # 8-byte Folded Spill
264; NOREMAT-NEXT:    add a2, a7, a2
265; NOREMAT-NEXT:    vle32.v v28, (a2)
266; NOREMAT-NEXT:    vle32.v v6, (a2)
267; NOREMAT-NEXT:    sf.vc.vv 3, 0, v30, v12
268; NOREMAT-NEXT:    addiw a2, a4, 1536
269; NOREMAT-NEXT:    sd a2, 480(sp) # 8-byte Folded Spill
270; NOREMAT-NEXT:    add a2, a7, a2
271; NOREMAT-NEXT:    vle32.v v12, (a2)
272; NOREMAT-NEXT:    vle32.v v30, (a2)
273; NOREMAT-NEXT:    slli a2, a5, 11
274; NOREMAT-NEXT:    sd a2, 472(sp) # 8-byte Folded Spill
275; NOREMAT-NEXT:    sf.vc.vv 3, 0, v4, v24
276; NOREMAT-NEXT:    add a2, a7, a2
277; NOREMAT-NEXT:    vle32.v v24, (a2)
278; NOREMAT-NEXT:    vle32.v v4, (a2)
279; NOREMAT-NEXT:    sf.vc.vv 3, 0, v22, v8
280; NOREMAT-NEXT:    lui a5, 5
281; NOREMAT-NEXT:    addiw a2, a5, -1536
282; NOREMAT-NEXT:    sd a2, 464(sp) # 8-byte Folded Spill
283; NOREMAT-NEXT:    add a2, a7, a2
284; NOREMAT-NEXT:    vle32.v v8, (a2)
285; NOREMAT-NEXT:    vle32.v v22, (a2)
286; NOREMAT-NEXT:    sf.vc.vv 3, 0, v26, v28
287; NOREMAT-NEXT:    slli a2, t2, 10
288; NOREMAT-NEXT:    sd a2, 456(sp) # 8-byte Folded Spill
289; NOREMAT-NEXT:    li t2, 19
290; NOREMAT-NEXT:    add a2, a7, a2
291; NOREMAT-NEXT:    vle32.v v26, (a2)
292; NOREMAT-NEXT:    vle32.v v28, (a2)
293; NOREMAT-NEXT:    sf.vc.vv 3, 0, v6, v12
294; NOREMAT-NEXT:    addiw a2, a5, -512
295; NOREMAT-NEXT:    sd a2, 448(sp) # 8-byte Folded Spill
296; NOREMAT-NEXT:    add a2, a7, a2
297; NOREMAT-NEXT:    vle32.v v12, (a2)
298; NOREMAT-NEXT:    vle32.v v6, (a2)
299; NOREMAT-NEXT:    sf.vc.vv 3, 0, v30, v24
300; NOREMAT-NEXT:    addiw a2, a5, 512
301; NOREMAT-NEXT:    sd a2, 440(sp) # 8-byte Folded Spill
302; NOREMAT-NEXT:    add a2, a7, a2
303; NOREMAT-NEXT:    vle32.v v24, (a2)
304; NOREMAT-NEXT:    vle32.v v30, (a2)
305; NOREMAT-NEXT:    slli a2, s7, 10
306; NOREMAT-NEXT:    sd a2, 432(sp) # 8-byte Folded Spill
307; NOREMAT-NEXT:    sf.vc.vv 3, 0, v4, v8
308; NOREMAT-NEXT:    add a2, a7, a2
309; NOREMAT-NEXT:    vle32.v v8, (a2)
310; NOREMAT-NEXT:    vle32.v v4, (a2)
311; NOREMAT-NEXT:    sf.vc.vv 3, 0, v22, v26
312; NOREMAT-NEXT:    addiw a2, a5, 1536
313; NOREMAT-NEXT:    sd a2, 424(sp) # 8-byte Folded Spill
314; NOREMAT-NEXT:    add a2, a7, a2
315; NOREMAT-NEXT:    vle32.v v22, (a2)
316; NOREMAT-NEXT:    vle32.v v26, (a2)
317; NOREMAT-NEXT:    slli a2, a6, 11
318; NOREMAT-NEXT:    sd a2, 416(sp) # 8-byte Folded Spill
319; NOREMAT-NEXT:    sf.vc.vv 3, 0, v28, v12
320; NOREMAT-NEXT:    add a2, a7, a2
321; NOREMAT-NEXT:    vle32.v v12, (a2)
322; NOREMAT-NEXT:    vle32.v v28, (a2)
323; NOREMAT-NEXT:    sf.vc.vv 3, 0, v6, v18
324; NOREMAT-NEXT:    lui a6, 6
325; NOREMAT-NEXT:    addiw a2, a6, -1536
326; NOREMAT-NEXT:    sd a2, 408(sp) # 8-byte Folded Spill
327; NOREMAT-NEXT:    add a2, a7, a2
328; NOREMAT-NEXT:    vle32.v v18, (a2)
329; NOREMAT-NEXT:    vle32.v v6, (a2)
330; NOREMAT-NEXT:    slli a2, s3, 10
331; NOREMAT-NEXT:    sd a2, 400(sp) # 8-byte Folded Spill
332; NOREMAT-NEXT:    sf.vc.vv 3, 0, v16, v24
333; NOREMAT-NEXT:    add a2, a7, a2
334; NOREMAT-NEXT:    vle32.v v16, (a2)
335; NOREMAT-NEXT:    vle32.v v24, (a2)
336; NOREMAT-NEXT:    sf.vc.vv 3, 0, v30, v8
337; NOREMAT-NEXT:    addiw a2, a6, -512
338; NOREMAT-NEXT:    sd a2, 392(sp) # 8-byte Folded Spill
339; NOREMAT-NEXT:    add a2, a7, a2
340; NOREMAT-NEXT:    vle32.v v8, (a2)
341; NOREMAT-NEXT:    vle32.v v30, (a2)
342; NOREMAT-NEXT:    sf.vc.vv 3, 0, v4, v22
343; NOREMAT-NEXT:    addiw a2, a6, 512
344; NOREMAT-NEXT:    sd a2, 384(sp) # 8-byte Folded Spill
345; NOREMAT-NEXT:    add a2, a7, a2
346; NOREMAT-NEXT:    vle32.v v22, (a2)
347; NOREMAT-NEXT:    vle32.v v4, (a2)
348; NOREMAT-NEXT:    slli a2, s0, 10
349; NOREMAT-NEXT:    sd a2, 376(sp) # 8-byte Folded Spill
350; NOREMAT-NEXT:    sf.vc.vv 3, 0, v26, v12
351; NOREMAT-NEXT:    add a2, a7, a2
352; NOREMAT-NEXT:    vle32.v v26, (a2)
353; NOREMAT-NEXT:    vle32.v v2, (a2)
354; NOREMAT-NEXT:    sf.vc.vv 3, 0, v28, v18
355; NOREMAT-NEXT:    addiw a2, a6, 1536
356; NOREMAT-NEXT:    sd a2, 368(sp) # 8-byte Folded Spill
357; NOREMAT-NEXT:    add a2, a7, a2
358; NOREMAT-NEXT:    vle32.v v18, (a2)
359; NOREMAT-NEXT:    vle32.v v28, (a2)
360; NOREMAT-NEXT:    slli a2, t5, 11
361; NOREMAT-NEXT:    sd a2, 360(sp) # 8-byte Folded Spill
362; NOREMAT-NEXT:    sf.vc.vv 3, 0, v6, v16
363; NOREMAT-NEXT:    add a2, a7, a2
364; NOREMAT-NEXT:    vle32.v v16, (a2)
365; NOREMAT-NEXT:    vle32.v v6, (a2)
366; NOREMAT-NEXT:    sf.vc.vv 3, 0, v24, v8
367; NOREMAT-NEXT:    lui s0, 7
368; NOREMAT-NEXT:    addiw a2, s0, -1536
369; NOREMAT-NEXT:    sd a2, 352(sp) # 8-byte Folded Spill
370; NOREMAT-NEXT:    add a2, a7, a2
371; NOREMAT-NEXT:    vle32.v v8, (a2)
372; NOREMAT-NEXT:    vle32.v v24, (a2)
373; NOREMAT-NEXT:    slli a2, t3, 10
374; NOREMAT-NEXT:    sd a2, 344(sp) # 8-byte Folded Spill
375; NOREMAT-NEXT:    sf.vc.vv 3, 0, v30, v14
376; NOREMAT-NEXT:    add a2, a7, a2
377; NOREMAT-NEXT:    vle32.v v14, (a2)
378; NOREMAT-NEXT:    vle32.v v30, (a2)
379; NOREMAT-NEXT:    addi a0, sp, 640
380; NOREMAT-NEXT:    vl2r.v v12, (a0) # Unknown-size Folded Reload
381; NOREMAT-NEXT:    sf.vc.vv 3, 0, v12, v22
382; NOREMAT-NEXT:    addiw a2, s0, -512
383; NOREMAT-NEXT:    sd a2, 336(sp) # 8-byte Folded Spill
384; NOREMAT-NEXT:    add a2, a7, a2
385; NOREMAT-NEXT:    vle32.v v22, (a2)
386; NOREMAT-NEXT:    vle32.v v12, (a2)
387; NOREMAT-NEXT:    sf.vc.vv 3, 0, v4, v26
388; NOREMAT-NEXT:    addiw a2, s0, 512
389; NOREMAT-NEXT:    sd a2, 328(sp) # 8-byte Folded Spill
390; NOREMAT-NEXT:    lui t3, 7
391; NOREMAT-NEXT:    add a2, a7, a2
392; NOREMAT-NEXT:    vle32.v v26, (a2)
393; NOREMAT-NEXT:    vle32.v v4, (a2)
394; NOREMAT-NEXT:    slli a2, t0, 10
395; NOREMAT-NEXT:    sd a2, 320(sp) # 8-byte Folded Spill
396; NOREMAT-NEXT:    sf.vc.vv 3, 0, v2, v18
397; NOREMAT-NEXT:    add a2, a7, a2
398; NOREMAT-NEXT:    vle32.v v18, (a2)
399; NOREMAT-NEXT:    vle32.v v2, (a2)
400; NOREMAT-NEXT:    sf.vc.vv 3, 0, v28, v16
401; NOREMAT-NEXT:    addiw a2, t3, 1536
402; NOREMAT-NEXT:    sd a2, 312(sp) # 8-byte Folded Spill
403; NOREMAT-NEXT:    add a2, a7, a2
404; NOREMAT-NEXT:    vle32.v v16, (a2)
405; NOREMAT-NEXT:    vle32.v v28, (a2)
406; NOREMAT-NEXT:    slli a2, a3, 11
407; NOREMAT-NEXT:    sd a2, 304(sp) # 8-byte Folded Spill
408; NOREMAT-NEXT:    sf.vc.vv 3, 0, v6, v8
409; NOREMAT-NEXT:    add a2, a7, a2
410; NOREMAT-NEXT:    vle32.v v8, (a2)
411; NOREMAT-NEXT:    vle32.v v6, (a2)
412; NOREMAT-NEXT:    sf.vc.vv 3, 0, v24, v14
413; NOREMAT-NEXT:    addiw a2, t4, -1536
414; NOREMAT-NEXT:    sd a2, 296(sp) # 8-byte Folded Spill
415; NOREMAT-NEXT:    add a2, a7, a2
416; NOREMAT-NEXT:    vle32.v v14, (a2)
417; NOREMAT-NEXT:    vle32.v v24, (a2)
418; NOREMAT-NEXT:    slli a2, t1, 10
419; NOREMAT-NEXT:    sd a2, 288(sp) # 8-byte Folded Spill
420; NOREMAT-NEXT:    sf.vc.vv 3, 0, v30, v22
421; NOREMAT-NEXT:    add a2, a7, a2
422; NOREMAT-NEXT:    vle32.v v22, (a2)
423; NOREMAT-NEXT:    vle32.v v30, (a2)
424; NOREMAT-NEXT:    addiw a0, t4, -512
425; NOREMAT-NEXT:    sd a0, 280(sp) # 8-byte Folded Spill
426; NOREMAT-NEXT:    add a0, a7, a0
427; NOREMAT-NEXT:    sf.vc.vv 3, 0, v12, v0
428; NOREMAT-NEXT:    vle32.v v12, (a0)
429; NOREMAT-NEXT:    vle32.v v0, (a0)
430; NOREMAT-NEXT:    sf.vc.vv 3, 0, v10, v26
431; NOREMAT-NEXT:    sf.vc.vv 3, 0, v4, v18
432; NOREMAT-NEXT:    sf.vc.vv 3, 0, v2, v16
433; NOREMAT-NEXT:    sf.vc.vv 3, 0, v28, v8
434; NOREMAT-NEXT:    sf.vc.vv 3, 0, v6, v14
435; NOREMAT-NEXT:    sf.vc.vv 3, 0, v24, v22
436; NOREMAT-NEXT:    sf.vc.vv 3, 0, v30, v12
437; NOREMAT-NEXT:    sf.vc.vv 3, 0, v0, v20
438; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
439; NOREMAT-NEXT:    addi a0, a1, 1024
440; NOREMAT-NEXT:    vse32.v v8, (a0)
441; NOREMAT-NEXT:    add s11, a1, s11
442; NOREMAT-NEXT:    sd s11, 272(sp) # 8-byte Folded Spill
443; NOREMAT-NEXT:    lui a0, 2
444; NOREMAT-NEXT:    add a0, a1, a0
445; NOREMAT-NEXT:    sd a0, 264(sp) # 8-byte Folded Spill
446; NOREMAT-NEXT:    lui a0, 3
447; NOREMAT-NEXT:    add a0, a1, a0
448; NOREMAT-NEXT:    sd a0, 256(sp) # 8-byte Folded Spill
449; NOREMAT-NEXT:    add a4, a1, a4
450; NOREMAT-NEXT:    sd a4, 248(sp) # 8-byte Folded Spill
451; NOREMAT-NEXT:    add a5, a1, a5
452; NOREMAT-NEXT:    sd a5, 240(sp) # 8-byte Folded Spill
453; NOREMAT-NEXT:    add a6, a1, a6
454; NOREMAT-NEXT:    sd a6, 232(sp) # 8-byte Folded Spill
455; NOREMAT-NEXT:    add t3, a1, t3
456; NOREMAT-NEXT:    sd t3, 224(sp) # 8-byte Folded Spill
457; NOREMAT-NEXT:    add a0, a1, t4
458; NOREMAT-NEXT:    sd a0, 216(sp) # 8-byte Folded Spill
459; NOREMAT-NEXT:    addiw a0, t4, 512
460; NOREMAT-NEXT:    sd a0, 192(sp) # 8-byte Folded Spill
461; NOREMAT-NEXT:    addiw a0, t4, 1024
462; NOREMAT-NEXT:    sd a0, 176(sp) # 8-byte Folded Spill
463; NOREMAT-NEXT:    addiw a0, t4, 1536
464; NOREMAT-NEXT:    sd a0, 160(sp) # 8-byte Folded Spill
465; NOREMAT-NEXT:    slli s1, s1, 11
466; NOREMAT-NEXT:    sd s1, 128(sp) # 8-byte Folded Spill
467; NOREMAT-NEXT:    lui a0, 9
468; NOREMAT-NEXT:    addiw a2, a0, -1536
469; NOREMAT-NEXT:    sd a2, 88(sp) # 8-byte Folded Spill
470; NOREMAT-NEXT:    addiw a2, a0, -1024
471; NOREMAT-NEXT:    sd a2, 72(sp) # 8-byte Folded Spill
472; NOREMAT-NEXT:    addiw a2, a0, -512
473; NOREMAT-NEXT:    sd a2, 40(sp) # 8-byte Folded Spill
474; NOREMAT-NEXT:    add a2, a1, a0
475; NOREMAT-NEXT:    sd a2, 208(sp) # 8-byte Folded Spill
476; NOREMAT-NEXT:    addiw s11, a0, 512
477; NOREMAT-NEXT:    addiw s7, a0, 1024
478; NOREMAT-NEXT:    addiw s3, a0, 1536
479; NOREMAT-NEXT:    slli s1, t2, 11
480; NOREMAT-NEXT:    lui a0, 10
481; NOREMAT-NEXT:    addiw t2, a0, -1536
482; NOREMAT-NEXT:    addiw a7, a0, -1024
483; NOREMAT-NEXT:    addiw a4, a0, -512
484; NOREMAT-NEXT:    add a2, a1, a0
485; NOREMAT-NEXT:    sd a2, 200(sp) # 8-byte Folded Spill
486; NOREMAT-NEXT:    addiw a0, a0, 512
487; NOREMAT-NEXT:    ld a2, 512(sp) # 8-byte Folded Reload
488; NOREMAT-NEXT:    add a2, a1, a2
489; NOREMAT-NEXT:    ld a3, 504(sp) # 8-byte Folded Reload
490; NOREMAT-NEXT:    add a3, a1, a3
491; NOREMAT-NEXT:    add a5, a1, t6
492; NOREMAT-NEXT:    add a6, a1, s2
493; NOREMAT-NEXT:    add t0, a1, s4
494; NOREMAT-NEXT:    add t1, a1, s5
495; NOREMAT-NEXT:    add t3, a1, s6
496; NOREMAT-NEXT:    add t4, a1, s8
497; NOREMAT-NEXT:    add t5, a1, s9
498; NOREMAT-NEXT:    add t6, a1, s10
499; NOREMAT-NEXT:    add s0, a1, ra
500; NOREMAT-NEXT:    ld s2, 624(sp) # 8-byte Folded Reload
501; NOREMAT-NEXT:    add s2, a1, s2
502; NOREMAT-NEXT:    ld s4, 616(sp) # 8-byte Folded Reload
503; NOREMAT-NEXT:    add s4, a1, s4
504; NOREMAT-NEXT:    ld s5, 608(sp) # 8-byte Folded Reload
505; NOREMAT-NEXT:    add s5, a1, s5
506; NOREMAT-NEXT:    ld s6, 600(sp) # 8-byte Folded Reload
507; NOREMAT-NEXT:    add s6, a1, s6
508; NOREMAT-NEXT:    ld s8, 592(sp) # 8-byte Folded Reload
509; NOREMAT-NEXT:    add s8, a1, s8
510; NOREMAT-NEXT:    ld s9, 584(sp) # 8-byte Folded Reload
511; NOREMAT-NEXT:    add s9, a1, s9
512; NOREMAT-NEXT:    ld s10, 576(sp) # 8-byte Folded Reload
513; NOREMAT-NEXT:    add s10, a1, s10
514; NOREMAT-NEXT:    ld ra, 568(sp) # 8-byte Folded Reload
515; NOREMAT-NEXT:    add ra, a1, ra
516; NOREMAT-NEXT:    sd ra, 16(sp) # 8-byte Folded Spill
517; NOREMAT-NEXT:    ld ra, 560(sp) # 8-byte Folded Reload
518; NOREMAT-NEXT:    add ra, a1, ra
519; NOREMAT-NEXT:    sd ra, 24(sp) # 8-byte Folded Spill
520; NOREMAT-NEXT:    ld ra, 552(sp) # 8-byte Folded Reload
521; NOREMAT-NEXT:    add ra, a1, ra
522; NOREMAT-NEXT:    sd ra, 32(sp) # 8-byte Folded Spill
523; NOREMAT-NEXT:    ld ra, 544(sp) # 8-byte Folded Reload
524; NOREMAT-NEXT:    add ra, a1, ra
525; NOREMAT-NEXT:    sd ra, 48(sp) # 8-byte Folded Spill
526; NOREMAT-NEXT:    ld ra, 536(sp) # 8-byte Folded Reload
527; NOREMAT-NEXT:    add ra, a1, ra
528; NOREMAT-NEXT:    sd ra, 56(sp) # 8-byte Folded Spill
529; NOREMAT-NEXT:    ld ra, 528(sp) # 8-byte Folded Reload
530; NOREMAT-NEXT:    add ra, a1, ra
531; NOREMAT-NEXT:    sd ra, 64(sp) # 8-byte Folded Spill
532; NOREMAT-NEXT:    ld ra, 520(sp) # 8-byte Folded Reload
533; NOREMAT-NEXT:    add ra, a1, ra
534; NOREMAT-NEXT:    sd ra, 80(sp) # 8-byte Folded Spill
535; NOREMAT-NEXT:    ld ra, 496(sp) # 8-byte Folded Reload
536; NOREMAT-NEXT:    add ra, a1, ra
537; NOREMAT-NEXT:    sd ra, 96(sp) # 8-byte Folded Spill
538; NOREMAT-NEXT:    ld ra, 488(sp) # 8-byte Folded Reload
539; NOREMAT-NEXT:    add ra, a1, ra
540; NOREMAT-NEXT:    sd ra, 104(sp) # 8-byte Folded Spill
541; NOREMAT-NEXT:    ld ra, 480(sp) # 8-byte Folded Reload
542; NOREMAT-NEXT:    add ra, a1, ra
543; NOREMAT-NEXT:    sd ra, 112(sp) # 8-byte Folded Spill
544; NOREMAT-NEXT:    ld ra, 472(sp) # 8-byte Folded Reload
545; NOREMAT-NEXT:    add ra, a1, ra
546; NOREMAT-NEXT:    sd ra, 120(sp) # 8-byte Folded Spill
547; NOREMAT-NEXT:    ld ra, 464(sp) # 8-byte Folded Reload
548; NOREMAT-NEXT:    add ra, a1, ra
549; NOREMAT-NEXT:    sd ra, 136(sp) # 8-byte Folded Spill
550; NOREMAT-NEXT:    ld ra, 456(sp) # 8-byte Folded Reload
551; NOREMAT-NEXT:    add ra, a1, ra
552; NOREMAT-NEXT:    sd ra, 144(sp) # 8-byte Folded Spill
553; NOREMAT-NEXT:    ld ra, 448(sp) # 8-byte Folded Reload
554; NOREMAT-NEXT:    add ra, a1, ra
555; NOREMAT-NEXT:    sd ra, 152(sp) # 8-byte Folded Spill
556; NOREMAT-NEXT:    ld ra, 440(sp) # 8-byte Folded Reload
557; NOREMAT-NEXT:    add ra, a1, ra
558; NOREMAT-NEXT:    sd ra, 168(sp) # 8-byte Folded Spill
559; NOREMAT-NEXT:    ld ra, 432(sp) # 8-byte Folded Reload
560; NOREMAT-NEXT:    add ra, a1, ra
561; NOREMAT-NEXT:    sd ra, 184(sp) # 8-byte Folded Spill
562; NOREMAT-NEXT:    ld ra, 424(sp) # 8-byte Folded Reload
563; NOREMAT-NEXT:    add ra, a1, ra
564; NOREMAT-NEXT:    sd ra, 424(sp) # 8-byte Folded Spill
565; NOREMAT-NEXT:    ld ra, 416(sp) # 8-byte Folded Reload
566; NOREMAT-NEXT:    add ra, a1, ra
567; NOREMAT-NEXT:    sd ra, 432(sp) # 8-byte Folded Spill
568; NOREMAT-NEXT:    ld ra, 408(sp) # 8-byte Folded Reload
569; NOREMAT-NEXT:    add ra, a1, ra
570; NOREMAT-NEXT:    sd ra, 440(sp) # 8-byte Folded Spill
571; NOREMAT-NEXT:    ld ra, 400(sp) # 8-byte Folded Reload
572; NOREMAT-NEXT:    add ra, a1, ra
573; NOREMAT-NEXT:    sd ra, 448(sp) # 8-byte Folded Spill
574; NOREMAT-NEXT:    ld ra, 392(sp) # 8-byte Folded Reload
575; NOREMAT-NEXT:    add ra, a1, ra
576; NOREMAT-NEXT:    sd ra, 456(sp) # 8-byte Folded Spill
577; NOREMAT-NEXT:    ld ra, 384(sp) # 8-byte Folded Reload
578; NOREMAT-NEXT:    add ra, a1, ra
579; NOREMAT-NEXT:    sd ra, 464(sp) # 8-byte Folded Spill
580; NOREMAT-NEXT:    ld ra, 376(sp) # 8-byte Folded Reload
581; NOREMAT-NEXT:    add ra, a1, ra
582; NOREMAT-NEXT:    sd ra, 472(sp) # 8-byte Folded Spill
583; NOREMAT-NEXT:    ld ra, 368(sp) # 8-byte Folded Reload
584; NOREMAT-NEXT:    add ra, a1, ra
585; NOREMAT-NEXT:    sd ra, 480(sp) # 8-byte Folded Spill
586; NOREMAT-NEXT:    ld ra, 360(sp) # 8-byte Folded Reload
587; NOREMAT-NEXT:    add ra, a1, ra
588; NOREMAT-NEXT:    sd ra, 488(sp) # 8-byte Folded Spill
589; NOREMAT-NEXT:    ld ra, 352(sp) # 8-byte Folded Reload
590; NOREMAT-NEXT:    add ra, a1, ra
591; NOREMAT-NEXT:    sd ra, 496(sp) # 8-byte Folded Spill
592; NOREMAT-NEXT:    ld ra, 344(sp) # 8-byte Folded Reload
593; NOREMAT-NEXT:    add ra, a1, ra
594; NOREMAT-NEXT:    sd ra, 504(sp) # 8-byte Folded Spill
595; NOREMAT-NEXT:    ld ra, 336(sp) # 8-byte Folded Reload
596; NOREMAT-NEXT:    add ra, a1, ra
597; NOREMAT-NEXT:    sd ra, 512(sp) # 8-byte Folded Spill
598; NOREMAT-NEXT:    ld ra, 328(sp) # 8-byte Folded Reload
599; NOREMAT-NEXT:    add ra, a1, ra
600; NOREMAT-NEXT:    sd ra, 520(sp) # 8-byte Folded Spill
601; NOREMAT-NEXT:    ld ra, 320(sp) # 8-byte Folded Reload
602; NOREMAT-NEXT:    add ra, a1, ra
603; NOREMAT-NEXT:    sd ra, 528(sp) # 8-byte Folded Spill
604; NOREMAT-NEXT:    ld ra, 312(sp) # 8-byte Folded Reload
605; NOREMAT-NEXT:    add ra, a1, ra
606; NOREMAT-NEXT:    sd ra, 536(sp) # 8-byte Folded Spill
607; NOREMAT-NEXT:    ld ra, 304(sp) # 8-byte Folded Reload
608; NOREMAT-NEXT:    add ra, a1, ra
609; NOREMAT-NEXT:    sd ra, 544(sp) # 8-byte Folded Spill
610; NOREMAT-NEXT:    ld ra, 296(sp) # 8-byte Folded Reload
611; NOREMAT-NEXT:    add ra, a1, ra
612; NOREMAT-NEXT:    sd ra, 552(sp) # 8-byte Folded Spill
613; NOREMAT-NEXT:    ld ra, 288(sp) # 8-byte Folded Reload
614; NOREMAT-NEXT:    add ra, a1, ra
615; NOREMAT-NEXT:    sd ra, 560(sp) # 8-byte Folded Spill
616; NOREMAT-NEXT:    ld ra, 280(sp) # 8-byte Folded Reload
617; NOREMAT-NEXT:    add ra, a1, ra
618; NOREMAT-NEXT:    sd ra, 568(sp) # 8-byte Folded Spill
619; NOREMAT-NEXT:    ld ra, 192(sp) # 8-byte Folded Reload
620; NOREMAT-NEXT:    add ra, a1, ra
621; NOREMAT-NEXT:    sd ra, 576(sp) # 8-byte Folded Spill
622; NOREMAT-NEXT:    ld ra, 176(sp) # 8-byte Folded Reload
623; NOREMAT-NEXT:    add ra, a1, ra
624; NOREMAT-NEXT:    sd ra, 584(sp) # 8-byte Folded Spill
625; NOREMAT-NEXT:    ld ra, 160(sp) # 8-byte Folded Reload
626; NOREMAT-NEXT:    add ra, a1, ra
627; NOREMAT-NEXT:    sd ra, 592(sp) # 8-byte Folded Spill
628; NOREMAT-NEXT:    ld ra, 128(sp) # 8-byte Folded Reload
629; NOREMAT-NEXT:    add ra, a1, ra
630; NOREMAT-NEXT:    sd ra, 600(sp) # 8-byte Folded Spill
631; NOREMAT-NEXT:    ld ra, 88(sp) # 8-byte Folded Reload
632; NOREMAT-NEXT:    add ra, a1, ra
633; NOREMAT-NEXT:    sd ra, 608(sp) # 8-byte Folded Spill
634; NOREMAT-NEXT:    ld ra, 72(sp) # 8-byte Folded Reload
635; NOREMAT-NEXT:    add ra, a1, ra
636; NOREMAT-NEXT:    sd ra, 616(sp) # 8-byte Folded Spill
637; NOREMAT-NEXT:    ld ra, 40(sp) # 8-byte Folded Reload
638; NOREMAT-NEXT:    add ra, a1, ra
639; NOREMAT-NEXT:    sd ra, 624(sp) # 8-byte Folded Spill
640; NOREMAT-NEXT:    add ra, a1, s11
641; NOREMAT-NEXT:    add s11, a1, s7
642; NOREMAT-NEXT:    add s7, a1, s3
643; NOREMAT-NEXT:    add s3, a1, s1
644; NOREMAT-NEXT:    add s1, a1, t2
645; NOREMAT-NEXT:    add t2, a1, a7
646; NOREMAT-NEXT:    add a7, a1, a4
647; NOREMAT-NEXT:    add a4, a1, a0
648; NOREMAT-NEXT:    addi a0, a1, 1536
649; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
650; NOREMAT-NEXT:    vse32.v v8, (a0)
651; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
652; NOREMAT-NEXT:    vse32.v v8, (a2)
653; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
654; NOREMAT-NEXT:    vse32.v v8, (a3)
655; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
656; NOREMAT-NEXT:    vse32.v v8, (a5)
657; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
658; NOREMAT-NEXT:    vse32.v v8, (a6)
659; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
660; NOREMAT-NEXT:    ld a0, 272(sp) # 8-byte Folded Reload
661; NOREMAT-NEXT:    vse32.v v8, (a0)
662; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
663; NOREMAT-NEXT:    vse32.v v8, (t0)
664; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
665; NOREMAT-NEXT:    vse32.v v8, (t1)
666; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
667; NOREMAT-NEXT:    vse32.v v8, (t3)
668; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
669; NOREMAT-NEXT:    vse32.v v8, (t4)
670; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
671; NOREMAT-NEXT:    vse32.v v8, (t5)
672; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
673; NOREMAT-NEXT:    vse32.v v8, (t6)
674; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
675; NOREMAT-NEXT:    vse32.v v8, (s0)
676; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
677; NOREMAT-NEXT:    ld a0, 264(sp) # 8-byte Folded Reload
678; NOREMAT-NEXT:    vse32.v v8, (a0)
679; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
680; NOREMAT-NEXT:    vse32.v v8, (s2)
681; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
682; NOREMAT-NEXT:    vse32.v v8, (s4)
683; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
684; NOREMAT-NEXT:    vse32.v v8, (s5)
685; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
686; NOREMAT-NEXT:    vse32.v v8, (s6)
687; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
688; NOREMAT-NEXT:    vse32.v v8, (s8)
689; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
690; NOREMAT-NEXT:    vse32.v v8, (s9)
691; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
692; NOREMAT-NEXT:    vse32.v v8, (s10)
693; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
694; NOREMAT-NEXT:    ld a0, 256(sp) # 8-byte Folded Reload
695; NOREMAT-NEXT:    vse32.v v8, (a0)
696; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
697; NOREMAT-NEXT:    ld a0, 16(sp) # 8-byte Folded Reload
698; NOREMAT-NEXT:    vse32.v v8, (a0)
699; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
700; NOREMAT-NEXT:    ld a0, 24(sp) # 8-byte Folded Reload
701; NOREMAT-NEXT:    vse32.v v8, (a0)
702; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
703; NOREMAT-NEXT:    ld a0, 32(sp) # 8-byte Folded Reload
704; NOREMAT-NEXT:    vse32.v v8, (a0)
705; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
706; NOREMAT-NEXT:    ld a0, 48(sp) # 8-byte Folded Reload
707; NOREMAT-NEXT:    vse32.v v8, (a0)
708; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
709; NOREMAT-NEXT:    ld a0, 56(sp) # 8-byte Folded Reload
710; NOREMAT-NEXT:    vse32.v v8, (a0)
711; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
712; NOREMAT-NEXT:    ld a0, 64(sp) # 8-byte Folded Reload
713; NOREMAT-NEXT:    vse32.v v8, (a0)
714; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
715; NOREMAT-NEXT:    ld a0, 80(sp) # 8-byte Folded Reload
716; NOREMAT-NEXT:    vse32.v v8, (a0)
717; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
718; NOREMAT-NEXT:    ld a0, 248(sp) # 8-byte Folded Reload
719; NOREMAT-NEXT:    vse32.v v8, (a0)
720; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
721; NOREMAT-NEXT:    ld a0, 96(sp) # 8-byte Folded Reload
722; NOREMAT-NEXT:    vse32.v v8, (a0)
723; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
724; NOREMAT-NEXT:    ld a0, 104(sp) # 8-byte Folded Reload
725; NOREMAT-NEXT:    vse32.v v8, (a0)
726; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
727; NOREMAT-NEXT:    ld a0, 112(sp) # 8-byte Folded Reload
728; NOREMAT-NEXT:    vse32.v v8, (a0)
729; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
730; NOREMAT-NEXT:    ld a0, 120(sp) # 8-byte Folded Reload
731; NOREMAT-NEXT:    vse32.v v8, (a0)
732; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
733; NOREMAT-NEXT:    ld a0, 136(sp) # 8-byte Folded Reload
734; NOREMAT-NEXT:    vse32.v v8, (a0)
735; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
736; NOREMAT-NEXT:    ld a0, 144(sp) # 8-byte Folded Reload
737; NOREMAT-NEXT:    vse32.v v8, (a0)
738; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
739; NOREMAT-NEXT:    ld a0, 152(sp) # 8-byte Folded Reload
740; NOREMAT-NEXT:    vse32.v v8, (a0)
741; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
742; NOREMAT-NEXT:    ld a0, 240(sp) # 8-byte Folded Reload
743; NOREMAT-NEXT:    vse32.v v8, (a0)
744; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
745; NOREMAT-NEXT:    ld a0, 168(sp) # 8-byte Folded Reload
746; NOREMAT-NEXT:    vse32.v v8, (a0)
747; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
748; NOREMAT-NEXT:    ld a0, 184(sp) # 8-byte Folded Reload
749; NOREMAT-NEXT:    vse32.v v8, (a0)
750; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
751; NOREMAT-NEXT:    ld a0, 424(sp) # 8-byte Folded Reload
752; NOREMAT-NEXT:    vse32.v v8, (a0)
753; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
754; NOREMAT-NEXT:    ld a0, 432(sp) # 8-byte Folded Reload
755; NOREMAT-NEXT:    vse32.v v8, (a0)
756; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
757; NOREMAT-NEXT:    ld a0, 440(sp) # 8-byte Folded Reload
758; NOREMAT-NEXT:    vse32.v v8, (a0)
759; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
760; NOREMAT-NEXT:    ld a0, 448(sp) # 8-byte Folded Reload
761; NOREMAT-NEXT:    vse32.v v8, (a0)
762; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
763; NOREMAT-NEXT:    ld a0, 456(sp) # 8-byte Folded Reload
764; NOREMAT-NEXT:    vse32.v v8, (a0)
765; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
766; NOREMAT-NEXT:    ld a0, 232(sp) # 8-byte Folded Reload
767; NOREMAT-NEXT:    vse32.v v8, (a0)
768; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
769; NOREMAT-NEXT:    ld a0, 464(sp) # 8-byte Folded Reload
770; NOREMAT-NEXT:    vse32.v v8, (a0)
771; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
772; NOREMAT-NEXT:    ld a0, 472(sp) # 8-byte Folded Reload
773; NOREMAT-NEXT:    vse32.v v8, (a0)
774; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
775; NOREMAT-NEXT:    ld a0, 480(sp) # 8-byte Folded Reload
776; NOREMAT-NEXT:    vse32.v v8, (a0)
777; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
778; NOREMAT-NEXT:    ld a0, 488(sp) # 8-byte Folded Reload
779; NOREMAT-NEXT:    vse32.v v8, (a0)
780; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
781; NOREMAT-NEXT:    ld a0, 496(sp) # 8-byte Folded Reload
782; NOREMAT-NEXT:    vse32.v v8, (a0)
783; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
784; NOREMAT-NEXT:    ld a0, 504(sp) # 8-byte Folded Reload
785; NOREMAT-NEXT:    vse32.v v8, (a0)
786; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
787; NOREMAT-NEXT:    ld a0, 512(sp) # 8-byte Folded Reload
788; NOREMAT-NEXT:    vse32.v v8, (a0)
789; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
790; NOREMAT-NEXT:    ld a0, 224(sp) # 8-byte Folded Reload
791; NOREMAT-NEXT:    vse32.v v8, (a0)
792; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
793; NOREMAT-NEXT:    ld a0, 520(sp) # 8-byte Folded Reload
794; NOREMAT-NEXT:    vse32.v v8, (a0)
795; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
796; NOREMAT-NEXT:    ld a0, 528(sp) # 8-byte Folded Reload
797; NOREMAT-NEXT:    vse32.v v8, (a0)
798; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
799; NOREMAT-NEXT:    ld a0, 536(sp) # 8-byte Folded Reload
800; NOREMAT-NEXT:    vse32.v v8, (a0)
801; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
802; NOREMAT-NEXT:    ld a0, 544(sp) # 8-byte Folded Reload
803; NOREMAT-NEXT:    vse32.v v8, (a0)
804; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
805; NOREMAT-NEXT:    ld a0, 552(sp) # 8-byte Folded Reload
806; NOREMAT-NEXT:    vse32.v v8, (a0)
807; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
808; NOREMAT-NEXT:    ld a0, 560(sp) # 8-byte Folded Reload
809; NOREMAT-NEXT:    vse32.v v8, (a0)
810; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
811; NOREMAT-NEXT:    ld a0, 568(sp) # 8-byte Folded Reload
812; NOREMAT-NEXT:    vse32.v v8, (a0)
813; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
814; NOREMAT-NEXT:    ld a0, 216(sp) # 8-byte Folded Reload
815; NOREMAT-NEXT:    vse32.v v8, (a0)
816; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
817; NOREMAT-NEXT:    ld a0, 576(sp) # 8-byte Folded Reload
818; NOREMAT-NEXT:    vse32.v v8, (a0)
819; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
820; NOREMAT-NEXT:    ld a0, 584(sp) # 8-byte Folded Reload
821; NOREMAT-NEXT:    vse32.v v8, (a0)
822; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
823; NOREMAT-NEXT:    ld a0, 592(sp) # 8-byte Folded Reload
824; NOREMAT-NEXT:    vse32.v v8, (a0)
825; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
826; NOREMAT-NEXT:    ld a0, 600(sp) # 8-byte Folded Reload
827; NOREMAT-NEXT:    vse32.v v8, (a0)
828; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
829; NOREMAT-NEXT:    ld a0, 608(sp) # 8-byte Folded Reload
830; NOREMAT-NEXT:    vse32.v v8, (a0)
831; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
832; NOREMAT-NEXT:    ld a0, 616(sp) # 8-byte Folded Reload
833; NOREMAT-NEXT:    vse32.v v8, (a0)
834; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
835; NOREMAT-NEXT:    ld a0, 624(sp) # 8-byte Folded Reload
836; NOREMAT-NEXT:    vse32.v v8, (a0)
837; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
838; NOREMAT-NEXT:    ld a0, 208(sp) # 8-byte Folded Reload
839; NOREMAT-NEXT:    vse32.v v8, (a0)
840; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
841; NOREMAT-NEXT:    vse32.v v8, (ra)
842; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
843; NOREMAT-NEXT:    vse32.v v8, (s11)
844; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
845; NOREMAT-NEXT:    vse32.v v8, (s7)
846; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
847; NOREMAT-NEXT:    vse32.v v8, (s3)
848; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
849; NOREMAT-NEXT:    vse32.v v8, (s1)
850; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
851; NOREMAT-NEXT:    vse32.v v8, (t2)
852; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
853; NOREMAT-NEXT:    vse32.v v8, (a7)
854; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
855; NOREMAT-NEXT:    ld a0, 200(sp) # 8-byte Folded Reload
856; NOREMAT-NEXT:    vse32.v v8, (a0)
857; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
858; NOREMAT-NEXT:    vse32.v v8, (a4)
859; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
860; NOREMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
861; NOREMAT-NEXT:    csrr a0, vlenb
862; NOREMAT-NEXT:    slli a0, a0, 1
863; NOREMAT-NEXT:    add sp, sp, a0
864; NOREMAT-NEXT:    .cfi_def_cfa sp, 752
865; NOREMAT-NEXT:    ld ra, 744(sp) # 8-byte Folded Reload
866; NOREMAT-NEXT:    ld s0, 736(sp) # 8-byte Folded Reload
867; NOREMAT-NEXT:    ld s1, 728(sp) # 8-byte Folded Reload
868; NOREMAT-NEXT:    ld s2, 720(sp) # 8-byte Folded Reload
869; NOREMAT-NEXT:    ld s3, 712(sp) # 8-byte Folded Reload
870; NOREMAT-NEXT:    ld s4, 704(sp) # 8-byte Folded Reload
871; NOREMAT-NEXT:    ld s5, 696(sp) # 8-byte Folded Reload
872; NOREMAT-NEXT:    ld s6, 688(sp) # 8-byte Folded Reload
873; NOREMAT-NEXT:    ld s7, 680(sp) # 8-byte Folded Reload
874; NOREMAT-NEXT:    ld s8, 672(sp) # 8-byte Folded Reload
875; NOREMAT-NEXT:    ld s9, 664(sp) # 8-byte Folded Reload
876; NOREMAT-NEXT:    ld s10, 656(sp) # 8-byte Folded Reload
877; NOREMAT-NEXT:    ld s11, 648(sp) # 8-byte Folded Reload
878; NOREMAT-NEXT:    .cfi_restore ra
879; NOREMAT-NEXT:    .cfi_restore s0
880; NOREMAT-NEXT:    .cfi_restore s1
881; NOREMAT-NEXT:    .cfi_restore s2
882; NOREMAT-NEXT:    .cfi_restore s3
883; NOREMAT-NEXT:    .cfi_restore s4
884; NOREMAT-NEXT:    .cfi_restore s5
885; NOREMAT-NEXT:    .cfi_restore s6
886; NOREMAT-NEXT:    .cfi_restore s7
887; NOREMAT-NEXT:    .cfi_restore s8
888; NOREMAT-NEXT:    .cfi_restore s9
889; NOREMAT-NEXT:    .cfi_restore s10
890; NOREMAT-NEXT:    .cfi_restore s11
891; NOREMAT-NEXT:    addi sp, sp, 752
892; NOREMAT-NEXT:    .cfi_def_cfa_offset 0
893; NOREMAT-NEXT:    ret
894;
895; REMAT-LABEL: test:
896; REMAT:       # %bb.0:
897; REMAT-NEXT:    addi sp, sp, -544
898; REMAT-NEXT:    .cfi_def_cfa_offset 544
899; REMAT-NEXT:    sd ra, 536(sp) # 8-byte Folded Spill
900; REMAT-NEXT:    sd s0, 528(sp) # 8-byte Folded Spill
901; REMAT-NEXT:    sd s1, 520(sp) # 8-byte Folded Spill
902; REMAT-NEXT:    sd s2, 512(sp) # 8-byte Folded Spill
903; REMAT-NEXT:    sd s3, 504(sp) # 8-byte Folded Spill
904; REMAT-NEXT:    sd s4, 496(sp) # 8-byte Folded Spill
905; REMAT-NEXT:    sd s5, 488(sp) # 8-byte Folded Spill
906; REMAT-NEXT:    sd s6, 480(sp) # 8-byte Folded Spill
907; REMAT-NEXT:    sd s7, 472(sp) # 8-byte Folded Spill
908; REMAT-NEXT:    sd s8, 464(sp) # 8-byte Folded Spill
909; REMAT-NEXT:    sd s9, 456(sp) # 8-byte Folded Spill
910; REMAT-NEXT:    sd s10, 448(sp) # 8-byte Folded Spill
911; REMAT-NEXT:    sd s11, 440(sp) # 8-byte Folded Spill
912; REMAT-NEXT:    .cfi_offset ra, -8
913; REMAT-NEXT:    .cfi_offset s0, -16
914; REMAT-NEXT:    .cfi_offset s1, -24
915; REMAT-NEXT:    .cfi_offset s2, -32
916; REMAT-NEXT:    .cfi_offset s3, -40
917; REMAT-NEXT:    .cfi_offset s4, -48
918; REMAT-NEXT:    .cfi_offset s5, -56
919; REMAT-NEXT:    .cfi_offset s6, -64
920; REMAT-NEXT:    .cfi_offset s7, -72
921; REMAT-NEXT:    .cfi_offset s8, -80
922; REMAT-NEXT:    .cfi_offset s9, -88
923; REMAT-NEXT:    .cfi_offset s10, -96
924; REMAT-NEXT:    .cfi_offset s11, -104
925; REMAT-NEXT:    csrr a2, vlenb
926; REMAT-NEXT:    li a3, 18
927; REMAT-NEXT:    mul a2, a2, a3
928; REMAT-NEXT:    sub sp, sp, a2
929; REMAT-NEXT:    .cfi_escape 0x0f, 0x0e, 0x72, 0x00, 0x11, 0xa0, 0x04, 0x22, 0x11, 0x12, 0x92, 0xa2, 0x38, 0x00, 0x1e, 0x22 # sp + 544 + 18 * vlenb
930; REMAT-NEXT:    li a4, 32
931; REMAT-NEXT:    addi a5, a0, 512
932; REMAT-NEXT:    addi a3, a0, 1024
933; REMAT-NEXT:    addi a2, a0, 1536
934; REMAT-NEXT:    li a6, 1
935; REMAT-NEXT:    slli a6, a6, 11
936; REMAT-NEXT:    li a7, 5
937; REMAT-NEXT:    slli a7, a7, 9
938; REMAT-NEXT:    li t0, 3
939; REMAT-NEXT:    slli t0, t0, 10
940; REMAT-NEXT:    li t1, 7
941; REMAT-NEXT:    slli t1, t1, 9
942; REMAT-NEXT:    lui t2, 1
943; REMAT-NEXT:    li t3, 9
944; REMAT-NEXT:    slli t3, t3, 9
945; REMAT-NEXT:    li t4, 5
946; REMAT-NEXT:    slli t4, t4, 10
947; REMAT-NEXT:    li t5, 11
948; REMAT-NEXT:    slli t5, t5, 9
949; REMAT-NEXT:    li t6, 3
950; REMAT-NEXT:    slli t6, t6, 11
951; REMAT-NEXT:    li s0, 13
952; REMAT-NEXT:    slli s0, s0, 9
953; REMAT-NEXT:    li s1, 7
954; REMAT-NEXT:    slli s1, s1, 10
955; REMAT-NEXT:    li s2, 15
956; REMAT-NEXT:    slli s2, s2, 9
957; REMAT-NEXT:    lui s3, 2
958; REMAT-NEXT:    li s4, 17
959; REMAT-NEXT:    slli s4, s4, 9
960; REMAT-NEXT:    li s5, 9
961; REMAT-NEXT:    slli s5, s5, 10
962; REMAT-NEXT:    li s6, 19
963; REMAT-NEXT:    slli s6, s6, 9
964; REMAT-NEXT:    li s7, 5
965; REMAT-NEXT:    slli s7, s7, 11
966; REMAT-NEXT:    li s8, 21
967; REMAT-NEXT:    slli s8, s8, 9
968; REMAT-NEXT:    li s9, 11
969; REMAT-NEXT:    slli s9, s9, 10
970; REMAT-NEXT:    li s10, 23
971; REMAT-NEXT:    slli s10, s10, 9
972; REMAT-NEXT:    lui s11, 3
973; REMAT-NEXT:    li ra, 25
974; REMAT-NEXT:    slli ra, ra, 9
975; REMAT-NEXT:    vsetvli zero, a4, e32, m2, ta, ma
976; REMAT-NEXT:    vle32.v v8, (a5)
977; REMAT-NEXT:    li a4, 13
978; REMAT-NEXT:    slli a4, a4, 10
979; REMAT-NEXT:    vle32.v v10, (a3)
980; REMAT-NEXT:    vle32.v v12, (a3)
981; REMAT-NEXT:    li a3, 27
982; REMAT-NEXT:    slli a3, a3, 9
983; REMAT-NEXT:    vle32.v v14, (a2)
984; REMAT-NEXT:    vle32.v v16, (a2)
985; REMAT-NEXT:    add a2, a0, a6
986; REMAT-NEXT:    vle32.v v18, (a2)
987; REMAT-NEXT:    vle32.v v20, (a2)
988; REMAT-NEXT:    add a2, a0, a7
989; REMAT-NEXT:    vle32.v v22, (a2)
990; REMAT-NEXT:    vle32.v v24, (a2)
991; REMAT-NEXT:    add a2, a0, t0
992; REMAT-NEXT:    vle32.v v26, (a2)
993; REMAT-NEXT:    vle32.v v28, (a2)
994; REMAT-NEXT:    add a2, a0, t1
995; REMAT-NEXT:    vle32.v v30, (a2)
996; REMAT-NEXT:    vle32.v v6, (a2)
997; REMAT-NEXT:    csrr a2, vlenb
998; REMAT-NEXT:    slli a2, a2, 4
999; REMAT-NEXT:    add a2, sp, a2
1000; REMAT-NEXT:    addi a2, a2, 432
1001; REMAT-NEXT:    vs2r.v v6, (a2) # Unknown-size Folded Spill
1002; REMAT-NEXT:    add a2, a0, t2
1003; REMAT-NEXT:    vle32.v v4, (a0)
1004; REMAT-NEXT:    vle32.v v2, (a2)
1005; REMAT-NEXT:    vle32.v v6, (a2)
1006; REMAT-NEXT:    csrr a2, vlenb
1007; REMAT-NEXT:    li a5, 14
1008; REMAT-NEXT:    mul a2, a2, a5
1009; REMAT-NEXT:    add a2, sp, a2
1010; REMAT-NEXT:    addi a2, a2, 432
1011; REMAT-NEXT:    vs2r.v v6, (a2) # Unknown-size Folded Spill
1012; REMAT-NEXT:    add a2, a0, t3
1013; REMAT-NEXT:    sf.vc.vv 3, 0, v4, v8
1014; REMAT-NEXT:    vle32.v v4, (a2)
1015; REMAT-NEXT:    sf.vc.vv 3, 0, v8, v10
1016; REMAT-NEXT:    vle32.v v6, (a2)
1017; REMAT-NEXT:    add a2, a0, t4
1018; REMAT-NEXT:    vle32.v v10, (a2)
1019; REMAT-NEXT:    sf.vc.vv 3, 0, v12, v14
1020; REMAT-NEXT:    vle32.v v0, (a2)
1021; REMAT-NEXT:    add a2, a0, t5
1022; REMAT-NEXT:    vle32.v v14, (a2)
1023; REMAT-NEXT:    sf.vc.vv 3, 0, v16, v18
1024; REMAT-NEXT:    vle32.v v8, (a2)
1025; REMAT-NEXT:    csrr a2, vlenb
1026; REMAT-NEXT:    li a5, 12
1027; REMAT-NEXT:    mul a2, a2, a5
1028; REMAT-NEXT:    add a2, sp, a2
1029; REMAT-NEXT:    addi a2, a2, 432
1030; REMAT-NEXT:    vs2r.v v8, (a2) # Unknown-size Folded Spill
1031; REMAT-NEXT:    add a2, a0, t6
1032; REMAT-NEXT:    vle32.v v18, (a2)
1033; REMAT-NEXT:    sf.vc.vv 3, 0, v20, v22
1034; REMAT-NEXT:    vle32.v v20, (a2)
1035; REMAT-NEXT:    add a2, a0, s0
1036; REMAT-NEXT:    vle32.v v22, (a2)
1037; REMAT-NEXT:    sf.vc.vv 3, 0, v24, v26
1038; REMAT-NEXT:    vle32.v v24, (a2)
1039; REMAT-NEXT:    add a2, a0, s1
1040; REMAT-NEXT:    vle32.v v26, (a2)
1041; REMAT-NEXT:    sf.vc.vv 3, 0, v28, v30
1042; REMAT-NEXT:    vle32.v v28, (a2)
1043; REMAT-NEXT:    add a2, a0, s2
1044; REMAT-NEXT:    vle32.v v8, (a2)
1045; REMAT-NEXT:    csrr a5, vlenb
1046; REMAT-NEXT:    slli a5, a5, 4
1047; REMAT-NEXT:    add a5, sp, a5
1048; REMAT-NEXT:    addi a5, a5, 432
1049; REMAT-NEXT:    vl2r.v v12, (a5) # Unknown-size Folded Reload
1050; REMAT-NEXT:    sf.vc.vv 3, 0, v12, v2
1051; REMAT-NEXT:    vle32.v v2, (a2)
1052; REMAT-NEXT:    add a2, a0, s3
1053; REMAT-NEXT:    vle32.v v12, (a2)
1054; REMAT-NEXT:    csrr a5, vlenb
1055; REMAT-NEXT:    li a6, 14
1056; REMAT-NEXT:    mul a5, a5, a6
1057; REMAT-NEXT:    add a5, sp, a5
1058; REMAT-NEXT:    addi a5, a5, 432
1059; REMAT-NEXT:    vl2r.v v16, (a5) # Unknown-size Folded Reload
1060; REMAT-NEXT:    sf.vc.vv 3, 0, v16, v4
1061; REMAT-NEXT:    vle32.v v30, (a2)
1062; REMAT-NEXT:    add a2, a0, s4
1063; REMAT-NEXT:    vle32.v v16, (a2)
1064; REMAT-NEXT:    sf.vc.vv 3, 0, v6, v10
1065; REMAT-NEXT:    vle32.v v6, (a2)
1066; REMAT-NEXT:    add a2, a0, s5
1067; REMAT-NEXT:    vle32.v v10, (a2)
1068; REMAT-NEXT:    sf.vc.vv 3, 0, v0, v14
1069; REMAT-NEXT:    vle32.v v4, (a2)
1070; REMAT-NEXT:    add a2, a0, s6
1071; REMAT-NEXT:    vle32.v v14, (a2)
1072; REMAT-NEXT:    csrr a5, vlenb
1073; REMAT-NEXT:    li a6, 12
1074; REMAT-NEXT:    mul a5, a5, a6
1075; REMAT-NEXT:    add a5, sp, a5
1076; REMAT-NEXT:    addi a5, a5, 432
1077; REMAT-NEXT:    vl2r.v v0, (a5) # Unknown-size Folded Reload
1078; REMAT-NEXT:    sf.vc.vv 3, 0, v0, v18
1079; REMAT-NEXT:    vle32.v v0, (a2)
1080; REMAT-NEXT:    add a2, a0, s7
1081; REMAT-NEXT:    vle32.v v18, (a2)
1082; REMAT-NEXT:    sf.vc.vv 3, 0, v20, v22
1083; REMAT-NEXT:    vle32.v v22, (a2)
1084; REMAT-NEXT:    add a2, a0, s8
1085; REMAT-NEXT:    vle32.v v20, (a2)
1086; REMAT-NEXT:    sf.vc.vv 3, 0, v24, v26
1087; REMAT-NEXT:    vle32.v v26, (a2)
1088; REMAT-NEXT:    add a2, a0, s9
1089; REMAT-NEXT:    vle32.v v24, (a2)
1090; REMAT-NEXT:    sf.vc.vv 3, 0, v28, v8
1091; REMAT-NEXT:    vle32.v v28, (a2)
1092; REMAT-NEXT:    add a2, a0, s10
1093; REMAT-NEXT:    vle32.v v8, (a2)
1094; REMAT-NEXT:    sf.vc.vv 3, 0, v2, v12
1095; REMAT-NEXT:    vle32.v v12, (a2)
1096; REMAT-NEXT:    csrr a2, vlenb
1097; REMAT-NEXT:    slli a2, a2, 3
1098; REMAT-NEXT:    add a2, sp, a2
1099; REMAT-NEXT:    addi a2, a2, 432
1100; REMAT-NEXT:    vs2r.v v12, (a2) # Unknown-size Folded Spill
1101; REMAT-NEXT:    add a2, a0, s11
1102; REMAT-NEXT:    vle32.v v12, (a2)
1103; REMAT-NEXT:    sf.vc.vv 3, 0, v30, v16
1104; REMAT-NEXT:    vle32.v v16, (a2)
1105; REMAT-NEXT:    add a2, a0, ra
1106; REMAT-NEXT:    vle32.v v2, (a2)
1107; REMAT-NEXT:    sf.vc.vv 3, 0, v6, v10
1108; REMAT-NEXT:    vle32.v v10, (a2)
1109; REMAT-NEXT:    csrr a2, vlenb
1110; REMAT-NEXT:    slli a2, a2, 1
1111; REMAT-NEXT:    add a2, sp, a2
1112; REMAT-NEXT:    addi a2, a2, 432
1113; REMAT-NEXT:    vs2r.v v10, (a2) # Unknown-size Folded Spill
1114; REMAT-NEXT:    add a2, a0, a4
1115; REMAT-NEXT:    vle32.v v10, (a2)
1116; REMAT-NEXT:    sf.vc.vv 3, 0, v4, v14
1117; REMAT-NEXT:    vle32.v v14, (a2)
1118; REMAT-NEXT:    csrr a2, vlenb
1119; REMAT-NEXT:    slli a2, a2, 2
1120; REMAT-NEXT:    add a2, sp, a2
1121; REMAT-NEXT:    addi a2, a2, 432
1122; REMAT-NEXT:    vs2r.v v14, (a2) # Unknown-size Folded Spill
1123; REMAT-NEXT:    add a2, a0, a3
1124; REMAT-NEXT:    vle32.v v14, (a2)
1125; REMAT-NEXT:    sf.vc.vv 3, 0, v0, v18
1126; REMAT-NEXT:    vle32.v v18, (a2)
1127; REMAT-NEXT:    csrr a2, vlenb
1128; REMAT-NEXT:    slli a2, a2, 4
1129; REMAT-NEXT:    add a2, sp, a2
1130; REMAT-NEXT:    addi a2, a2, 432
1131; REMAT-NEXT:    vs2r.v v18, (a2) # Unknown-size Folded Spill
1132; REMAT-NEXT:    li a5, 7
1133; REMAT-NEXT:    slli a5, a5, 11
1134; REMAT-NEXT:    add a2, a0, a5
1135; REMAT-NEXT:    vle32.v v18, (a2)
1136; REMAT-NEXT:    addi a3, sp, 432
1137; REMAT-NEXT:    vs2r.v v18, (a3) # Unknown-size Folded Spill
1138; REMAT-NEXT:    sf.vc.vv 3, 0, v22, v20
1139; REMAT-NEXT:    vle32.v v18, (a2)
1140; REMAT-NEXT:    csrr a2, vlenb
1141; REMAT-NEXT:    li a3, 14
1142; REMAT-NEXT:    mul a2, a2, a3
1143; REMAT-NEXT:    add a2, sp, a2
1144; REMAT-NEXT:    addi a2, a2, 432
1145; REMAT-NEXT:    vs2r.v v18, (a2) # Unknown-size Folded Spill
1146; REMAT-NEXT:    li a2, 29
1147; REMAT-NEXT:    slli a2, a2, 9
1148; REMAT-NEXT:    add a2, a0, a2
1149; REMAT-NEXT:    vle32.v v18, (a2)
1150; REMAT-NEXT:    sf.vc.vv 3, 0, v26, v24
1151; REMAT-NEXT:    vle32.v v20, (a2)
1152; REMAT-NEXT:    csrr a2, vlenb
1153; REMAT-NEXT:    li a3, 12
1154; REMAT-NEXT:    mul a2, a2, a3
1155; REMAT-NEXT:    add a2, sp, a2
1156; REMAT-NEXT:    addi a2, a2, 432
1157; REMAT-NEXT:    vs2r.v v20, (a2) # Unknown-size Folded Spill
1158; REMAT-NEXT:    li a2, 15
1159; REMAT-NEXT:    slli a2, a2, 10
1160; REMAT-NEXT:    add a2, a0, a2
1161; REMAT-NEXT:    vle32.v v30, (a2)
1162; REMAT-NEXT:    sf.vc.vv 3, 0, v28, v8
1163; REMAT-NEXT:    vle32.v v8, (a2)
1164; REMAT-NEXT:    csrr a2, vlenb
1165; REMAT-NEXT:    li a3, 10
1166; REMAT-NEXT:    mul a2, a2, a3
1167; REMAT-NEXT:    add a2, sp, a2
1168; REMAT-NEXT:    addi a2, a2, 432
1169; REMAT-NEXT:    vs2r.v v8, (a2) # Unknown-size Folded Spill
1170; REMAT-NEXT:    li a2, 31
1171; REMAT-NEXT:    slli a2, a2, 9
1172; REMAT-NEXT:    add a2, a0, a2
1173; REMAT-NEXT:    vle32.v v6, (a2)
1174; REMAT-NEXT:    csrr a3, vlenb
1175; REMAT-NEXT:    slli a3, a3, 3
1176; REMAT-NEXT:    add a3, sp, a3
1177; REMAT-NEXT:    addi a3, a3, 432
1178; REMAT-NEXT:    vl2r.v v8, (a3) # Unknown-size Folded Reload
1179; REMAT-NEXT:    sf.vc.vv 3, 0, v8, v12
1180; REMAT-NEXT:    vle32.v v8, (a2)
1181; REMAT-NEXT:    csrr a2, vlenb
1182; REMAT-NEXT:    slli a2, a2, 3
1183; REMAT-NEXT:    add a2, sp, a2
1184; REMAT-NEXT:    addi a2, a2, 432
1185; REMAT-NEXT:    vs2r.v v8, (a2) # Unknown-size Folded Spill
1186; REMAT-NEXT:    lui a2, 4
1187; REMAT-NEXT:    add a2, a0, a2
1188; REMAT-NEXT:    vle32.v v4, (a2)
1189; REMAT-NEXT:    sf.vc.vv 3, 0, v16, v2
1190; REMAT-NEXT:    vle32.v v8, (a2)
1191; REMAT-NEXT:    csrr a2, vlenb
1192; REMAT-NEXT:    li a3, 6
1193; REMAT-NEXT:    mul a2, a2, a3
1194; REMAT-NEXT:    add a2, sp, a2
1195; REMAT-NEXT:    addi a2, a2, 432
1196; REMAT-NEXT:    vs2r.v v8, (a2) # Unknown-size Folded Spill
1197; REMAT-NEXT:    lui a2, 4
1198; REMAT-NEXT:    addiw a2, a2, 512
1199; REMAT-NEXT:    add a2, a0, a2
1200; REMAT-NEXT:    vle32.v v2, (a2)
1201; REMAT-NEXT:    csrr a3, vlenb
1202; REMAT-NEXT:    slli a3, a3, 1
1203; REMAT-NEXT:    add a3, sp, a3
1204; REMAT-NEXT:    addi a3, a3, 432
1205; REMAT-NEXT:    vl2r.v v8, (a3) # Unknown-size Folded Reload
1206; REMAT-NEXT:    sf.vc.vv 3, 0, v8, v10
1207; REMAT-NEXT:    vle32.v v20, (a2)
1208; REMAT-NEXT:    li a2, 17
1209; REMAT-NEXT:    slli a2, a2, 10
1210; REMAT-NEXT:    add a2, a0, a2
1211; REMAT-NEXT:    vle32.v v0, (a2)
1212; REMAT-NEXT:    csrr a3, vlenb
1213; REMAT-NEXT:    slli a3, a3, 2
1214; REMAT-NEXT:    add a3, sp, a3
1215; REMAT-NEXT:    addi a3, a3, 432
1216; REMAT-NEXT:    vl2r.v v8, (a3) # Unknown-size Folded Reload
1217; REMAT-NEXT:    sf.vc.vv 3, 0, v8, v14
1218; REMAT-NEXT:    vle32.v v22, (a2)
1219; REMAT-NEXT:    lui a2, 4
1220; REMAT-NEXT:    addiw a2, a2, 1536
1221; REMAT-NEXT:    add a2, a0, a2
1222; REMAT-NEXT:    vle32.v v24, (a2)
1223; REMAT-NEXT:    csrr a3, vlenb
1224; REMAT-NEXT:    slli a3, a3, 4
1225; REMAT-NEXT:    add a3, sp, a3
1226; REMAT-NEXT:    addi a3, a3, 432
1227; REMAT-NEXT:    vl2r.v v8, (a3) # Unknown-size Folded Reload
1228; REMAT-NEXT:    addi a3, sp, 432
1229; REMAT-NEXT:    vl2r.v v10, (a3) # Unknown-size Folded Reload
1230; REMAT-NEXT:    sf.vc.vv 3, 0, v8, v10
1231; REMAT-NEXT:    vle32.v v8, (a2)
1232; REMAT-NEXT:    li a2, 9
1233; REMAT-NEXT:    slli a2, a2, 11
1234; REMAT-NEXT:    add a2, a0, a2
1235; REMAT-NEXT:    vle32.v v26, (a2)
1236; REMAT-NEXT:    csrr a3, vlenb
1237; REMAT-NEXT:    li a4, 14
1238; REMAT-NEXT:    mul a3, a3, a4
1239; REMAT-NEXT:    add a3, sp, a3
1240; REMAT-NEXT:    addi a3, a3, 432
1241; REMAT-NEXT:    vl2r.v v10, (a3) # Unknown-size Folded Reload
1242; REMAT-NEXT:    sf.vc.vv 3, 0, v10, v18
1243; REMAT-NEXT:    vle32.v v10, (a2)
1244; REMAT-NEXT:    lui a2, 5
1245; REMAT-NEXT:    addiw a2, a2, -1536
1246; REMAT-NEXT:    add a2, a0, a2
1247; REMAT-NEXT:    vle32.v v28, (a2)
1248; REMAT-NEXT:    csrr a3, vlenb
1249; REMAT-NEXT:    li a4, 12
1250; REMAT-NEXT:    mul a3, a3, a4
1251; REMAT-NEXT:    add a3, sp, a3
1252; REMAT-NEXT:    addi a3, a3, 432
1253; REMAT-NEXT:    vl2r.v v12, (a3) # Unknown-size Folded Reload
1254; REMAT-NEXT:    sf.vc.vv 3, 0, v12, v30
1255; REMAT-NEXT:    vle32.v v12, (a2)
1256; REMAT-NEXT:    li a2, 19
1257; REMAT-NEXT:    slli a2, a2, 10
1258; REMAT-NEXT:    add a2, a0, a2
1259; REMAT-NEXT:    vle32.v v30, (a2)
1260; REMAT-NEXT:    csrr a3, vlenb
1261; REMAT-NEXT:    li a4, 10
1262; REMAT-NEXT:    mul a3, a3, a4
1263; REMAT-NEXT:    add a3, sp, a3
1264; REMAT-NEXT:    addi a3, a3, 432
1265; REMAT-NEXT:    vl2r.v v14, (a3) # Unknown-size Folded Reload
1266; REMAT-NEXT:    sf.vc.vv 3, 0, v14, v6
1267; REMAT-NEXT:    vle32.v v14, (a2)
1268; REMAT-NEXT:    lui a2, 5
1269; REMAT-NEXT:    addiw a2, a2, -512
1270; REMAT-NEXT:    add a2, a0, a2
1271; REMAT-NEXT:    vle32.v v6, (a2)
1272; REMAT-NEXT:    csrr a3, vlenb
1273; REMAT-NEXT:    slli a3, a3, 3
1274; REMAT-NEXT:    add a3, sp, a3
1275; REMAT-NEXT:    addi a3, a3, 432
1276; REMAT-NEXT:    vl2r.v v16, (a3) # Unknown-size Folded Reload
1277; REMAT-NEXT:    sf.vc.vv 3, 0, v16, v4
1278; REMAT-NEXT:    vle32.v v16, (a2)
1279; REMAT-NEXT:    lui a2, 5
1280; REMAT-NEXT:    add a2, a0, a2
1281; REMAT-NEXT:    vle32.v v4, (a2)
1282; REMAT-NEXT:    csrr a3, vlenb
1283; REMAT-NEXT:    li a4, 6
1284; REMAT-NEXT:    mul a3, a3, a4
1285; REMAT-NEXT:    add a3, sp, a3
1286; REMAT-NEXT:    addi a3, a3, 432
1287; REMAT-NEXT:    vl2r.v v18, (a3) # Unknown-size Folded Reload
1288; REMAT-NEXT:    sf.vc.vv 3, 0, v18, v2
1289; REMAT-NEXT:    vle32.v v18, (a2)
1290; REMAT-NEXT:    lui a2, 5
1291; REMAT-NEXT:    addiw a2, a2, 512
1292; REMAT-NEXT:    add a2, a0, a2
1293; REMAT-NEXT:    vle32.v v2, (a2)
1294; REMAT-NEXT:    sf.vc.vv 3, 0, v20, v0
1295; REMAT-NEXT:    vle32.v v20, (a2)
1296; REMAT-NEXT:    li s7, 21
1297; REMAT-NEXT:    slli s7, s7, 10
1298; REMAT-NEXT:    add a2, a0, s7
1299; REMAT-NEXT:    vle32.v v0, (a2)
1300; REMAT-NEXT:    sf.vc.vv 3, 0, v22, v24
1301; REMAT-NEXT:    vle32.v v22, (a2)
1302; REMAT-NEXT:    lui s4, 5
1303; REMAT-NEXT:    addiw s4, s4, 1536
1304; REMAT-NEXT:    add a2, a0, s4
1305; REMAT-NEXT:    vle32.v v24, (a2)
1306; REMAT-NEXT:    sf.vc.vv 3, 0, v8, v26
1307; REMAT-NEXT:    vle32.v v8, (a2)
1308; REMAT-NEXT:    li a2, 11
1309; REMAT-NEXT:    slli a2, a2, 11
1310; REMAT-NEXT:    add a2, a0, a2
1311; REMAT-NEXT:    vle32.v v26, (a2)
1312; REMAT-NEXT:    sf.vc.vv 3, 0, v10, v28
1313; REMAT-NEXT:    vle32.v v10, (a2)
1314; REMAT-NEXT:    lui s3, 6
1315; REMAT-NEXT:    addiw s3, s3, -1536
1316; REMAT-NEXT:    add a2, a0, s3
1317; REMAT-NEXT:    vle32.v v28, (a2)
1318; REMAT-NEXT:    sf.vc.vv 3, 0, v12, v30
1319; REMAT-NEXT:    vle32.v v12, (a2)
1320; REMAT-NEXT:    li s2, 23
1321; REMAT-NEXT:    slli s2, s2, 10
1322; REMAT-NEXT:    add a2, a0, s2
1323; REMAT-NEXT:    vle32.v v30, (a2)
1324; REMAT-NEXT:    sf.vc.vv 3, 0, v14, v6
1325; REMAT-NEXT:    vle32.v v14, (a2)
1326; REMAT-NEXT:    lui a2, 6
1327; REMAT-NEXT:    addiw a2, a2, -512
1328; REMAT-NEXT:    add a2, a0, a2
1329; REMAT-NEXT:    vle32.v v6, (a2)
1330; REMAT-NEXT:    sf.vc.vv 3, 0, v16, v4
1331; REMAT-NEXT:    vle32.v v16, (a2)
1332; REMAT-NEXT:    lui a2, 6
1333; REMAT-NEXT:    add a2, a0, a2
1334; REMAT-NEXT:    lui s1, 6
1335; REMAT-NEXT:    vle32.v v4, (a2)
1336; REMAT-NEXT:    sf.vc.vv 3, 0, v18, v2
1337; REMAT-NEXT:    vle32.v v18, (a2)
1338; REMAT-NEXT:    lui s0, 6
1339; REMAT-NEXT:    addiw s0, s0, 512
1340; REMAT-NEXT:    add a2, a0, s0
1341; REMAT-NEXT:    vle32.v v2, (a2)
1342; REMAT-NEXT:    sf.vc.vv 3, 0, v20, v0
1343; REMAT-NEXT:    vle32.v v20, (a2)
1344; REMAT-NEXT:    li a2, 25
1345; REMAT-NEXT:    slli a2, a2, 10
1346; REMAT-NEXT:    add a2, a0, a2
1347; REMAT-NEXT:    vle32.v v0, (a2)
1348; REMAT-NEXT:    sf.vc.vv 3, 0, v22, v24
1349; REMAT-NEXT:    vle32.v v22, (a2)
1350; REMAT-NEXT:    lui t6, 6
1351; REMAT-NEXT:    addiw t6, t6, 1536
1352; REMAT-NEXT:    add a2, a0, t6
1353; REMAT-NEXT:    vle32.v v24, (a2)
1354; REMAT-NEXT:    sf.vc.vv 3, 0, v8, v26
1355; REMAT-NEXT:    vle32.v v8, (a2)
1356; REMAT-NEXT:    li t5, 13
1357; REMAT-NEXT:    slli t5, t5, 11
1358; REMAT-NEXT:    add a2, a0, t5
1359; REMAT-NEXT:    vle32.v v26, (a2)
1360; REMAT-NEXT:    sf.vc.vv 3, 0, v10, v28
1361; REMAT-NEXT:    vle32.v v10, (a2)
1362; REMAT-NEXT:    lui a2, 7
1363; REMAT-NEXT:    addiw a2, a2, -1536
1364; REMAT-NEXT:    add a2, a0, a2
1365; REMAT-NEXT:    vle32.v v28, (a2)
1366; REMAT-NEXT:    sf.vc.vv 3, 0, v12, v30
1367; REMAT-NEXT:    vle32.v v12, (a2)
1368; REMAT-NEXT:    li t4, 27
1369; REMAT-NEXT:    slli t4, t4, 10
1370; REMAT-NEXT:    add a2, a0, t4
1371; REMAT-NEXT:    vle32.v v30, (a2)
1372; REMAT-NEXT:    sf.vc.vv 3, 0, v14, v6
1373; REMAT-NEXT:    vle32.v v14, (a2)
1374; REMAT-NEXT:    lui a2, 7
1375; REMAT-NEXT:    addiw a2, a2, -512
1376; REMAT-NEXT:    add a2, a0, a2
1377; REMAT-NEXT:    vle32.v v6, (a2)
1378; REMAT-NEXT:    sf.vc.vv 3, 0, v16, v4
1379; REMAT-NEXT:    vle32.v v16, (a2)
1380; REMAT-NEXT:    lui a2, 7
1381; REMAT-NEXT:    add a2, a0, a2
1382; REMAT-NEXT:    lui t3, 7
1383; REMAT-NEXT:    vle32.v v4, (a2)
1384; REMAT-NEXT:    sf.vc.vv 3, 0, v18, v2
1385; REMAT-NEXT:    vle32.v v18, (a2)
1386; REMAT-NEXT:    lui t2, 7
1387; REMAT-NEXT:    addiw t2, t2, 512
1388; REMAT-NEXT:    add a2, a0, t2
1389; REMAT-NEXT:    vle32.v v2, (a2)
1390; REMAT-NEXT:    sf.vc.vv 3, 0, v20, v0
1391; REMAT-NEXT:    vle32.v v20, (a2)
1392; REMAT-NEXT:    li t1, 29
1393; REMAT-NEXT:    slli t1, t1, 10
1394; REMAT-NEXT:    add a2, a0, t1
1395; REMAT-NEXT:    vle32.v v0, (a2)
1396; REMAT-NEXT:    sf.vc.vv 3, 0, v22, v24
1397; REMAT-NEXT:    vle32.v v22, (a2)
1398; REMAT-NEXT:    lui t0, 7
1399; REMAT-NEXT:    addiw t0, t0, 1536
1400; REMAT-NEXT:    add a2, a0, t0
1401; REMAT-NEXT:    vle32.v v24, (a2)
1402; REMAT-NEXT:    sf.vc.vv 3, 0, v8, v26
1403; REMAT-NEXT:    vle32.v v8, (a2)
1404; REMAT-NEXT:    li a7, 15
1405; REMAT-NEXT:    slli a7, a7, 11
1406; REMAT-NEXT:    add a2, a0, a7
1407; REMAT-NEXT:    vle32.v v26, (a2)
1408; REMAT-NEXT:    sf.vc.vv 3, 0, v10, v28
1409; REMAT-NEXT:    vle32.v v10, (a2)
1410; REMAT-NEXT:    lui a6, 8
1411; REMAT-NEXT:    addiw a6, a6, -1536
1412; REMAT-NEXT:    add a2, a0, a6
1413; REMAT-NEXT:    vle32.v v28, (a2)
1414; REMAT-NEXT:    sf.vc.vv 3, 0, v12, v30
1415; REMAT-NEXT:    vle32.v v12, (a2)
1416; REMAT-NEXT:    li a4, 31
1417; REMAT-NEXT:    slli a4, a4, 10
1418; REMAT-NEXT:    add a2, a0, a4
1419; REMAT-NEXT:    vle32.v v30, (a2)
1420; REMAT-NEXT:    sf.vc.vv 3, 0, v14, v6
1421; REMAT-NEXT:    vle32.v v14, (a2)
1422; REMAT-NEXT:    lui a3, 8
1423; REMAT-NEXT:    addiw a3, a3, -512
1424; REMAT-NEXT:    add a2, a0, a3
1425; REMAT-NEXT:    vle32.v v6, (a2)
1426; REMAT-NEXT:    sf.vc.vv 3, 0, v16, v4
1427; REMAT-NEXT:    vle32.v v16, (a2)
1428; REMAT-NEXT:    lui a2, 8
1429; REMAT-NEXT:    add a0, a0, a2
1430; REMAT-NEXT:    vle32.v v4, (a0)
1431; REMAT-NEXT:    sf.vc.vv 3, 0, v18, v2
1432; REMAT-NEXT:    sf.vc.vv 3, 0, v20, v0
1433; REMAT-NEXT:    sf.vc.vv 3, 0, v22, v24
1434; REMAT-NEXT:    sf.vc.vv 3, 0, v8, v26
1435; REMAT-NEXT:    sf.vc.vv 3, 0, v10, v28
1436; REMAT-NEXT:    sf.vc.vv 3, 0, v12, v30
1437; REMAT-NEXT:    sf.vc.vv 3, 0, v14, v6
1438; REMAT-NEXT:    sf.vc.vv 3, 0, v16, v4
1439; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1440; REMAT-NEXT:    addi a0, a1, 1024
1441; REMAT-NEXT:    vse32.v v8, (a0)
1442; REMAT-NEXT:    li a0, 1
1443; REMAT-NEXT:    slli a0, a0, 11
1444; REMAT-NEXT:    add a0, a1, a0
1445; REMAT-NEXT:    sd a0, 416(sp) # 8-byte Folded Spill
1446; REMAT-NEXT:    li a0, 5
1447; REMAT-NEXT:    slli a0, a0, 9
1448; REMAT-NEXT:    add a0, a1, a0
1449; REMAT-NEXT:    sd a0, 408(sp) # 8-byte Folded Spill
1450; REMAT-NEXT:    li a0, 3
1451; REMAT-NEXT:    slli a0, a0, 10
1452; REMAT-NEXT:    add a0, a1, a0
1453; REMAT-NEXT:    sd a0, 400(sp) # 8-byte Folded Spill
1454; REMAT-NEXT:    li a0, 7
1455; REMAT-NEXT:    slli a0, a0, 9
1456; REMAT-NEXT:    add a0, a1, a0
1457; REMAT-NEXT:    sd a0, 392(sp) # 8-byte Folded Spill
1458; REMAT-NEXT:    lui a0, 1
1459; REMAT-NEXT:    add a0, a1, a0
1460; REMAT-NEXT:    sd a0, 384(sp) # 8-byte Folded Spill
1461; REMAT-NEXT:    li a0, 9
1462; REMAT-NEXT:    slli a0, a0, 9
1463; REMAT-NEXT:    add a0, a1, a0
1464; REMAT-NEXT:    sd a0, 376(sp) # 8-byte Folded Spill
1465; REMAT-NEXT:    li a0, 5
1466; REMAT-NEXT:    slli a0, a0, 10
1467; REMAT-NEXT:    add a0, a1, a0
1468; REMAT-NEXT:    sd a0, 368(sp) # 8-byte Folded Spill
1469; REMAT-NEXT:    li a0, 11
1470; REMAT-NEXT:    slli a0, a0, 9
1471; REMAT-NEXT:    add a0, a1, a0
1472; REMAT-NEXT:    sd a0, 360(sp) # 8-byte Folded Spill
1473; REMAT-NEXT:    li a0, 3
1474; REMAT-NEXT:    slli a0, a0, 11
1475; REMAT-NEXT:    add a0, a1, a0
1476; REMAT-NEXT:    sd a0, 352(sp) # 8-byte Folded Spill
1477; REMAT-NEXT:    li a0, 13
1478; REMAT-NEXT:    slli a0, a0, 9
1479; REMAT-NEXT:    add a0, a1, a0
1480; REMAT-NEXT:    sd a0, 344(sp) # 8-byte Folded Spill
1481; REMAT-NEXT:    li a0, 7
1482; REMAT-NEXT:    slli a0, a0, 10
1483; REMAT-NEXT:    add a0, a1, a0
1484; REMAT-NEXT:    sd a0, 336(sp) # 8-byte Folded Spill
1485; REMAT-NEXT:    li a0, 15
1486; REMAT-NEXT:    slli a0, a0, 9
1487; REMAT-NEXT:    add a0, a1, a0
1488; REMAT-NEXT:    sd a0, 328(sp) # 8-byte Folded Spill
1489; REMAT-NEXT:    lui a0, 2
1490; REMAT-NEXT:    add a0, a1, a0
1491; REMAT-NEXT:    sd a0, 320(sp) # 8-byte Folded Spill
1492; REMAT-NEXT:    li a0, 17
1493; REMAT-NEXT:    slli a0, a0, 9
1494; REMAT-NEXT:    add a0, a1, a0
1495; REMAT-NEXT:    sd a0, 312(sp) # 8-byte Folded Spill
1496; REMAT-NEXT:    add s5, a1, s5
1497; REMAT-NEXT:    sd s5, 304(sp) # 8-byte Folded Spill
1498; REMAT-NEXT:    add s6, a1, s6
1499; REMAT-NEXT:    sd s6, 296(sp) # 8-byte Folded Spill
1500; REMAT-NEXT:    li a0, 5
1501; REMAT-NEXT:    slli a0, a0, 11
1502; REMAT-NEXT:    add a0, a1, a0
1503; REMAT-NEXT:    sd a0, 288(sp) # 8-byte Folded Spill
1504; REMAT-NEXT:    add s8, a1, s8
1505; REMAT-NEXT:    sd s8, 280(sp) # 8-byte Folded Spill
1506; REMAT-NEXT:    add s9, a1, s9
1507; REMAT-NEXT:    sd s9, 272(sp) # 8-byte Folded Spill
1508; REMAT-NEXT:    add s10, a1, s10
1509; REMAT-NEXT:    sd s10, 264(sp) # 8-byte Folded Spill
1510; REMAT-NEXT:    add s11, a1, s11
1511; REMAT-NEXT:    sd s11, 256(sp) # 8-byte Folded Spill
1512; REMAT-NEXT:    add ra, a1, ra
1513; REMAT-NEXT:    sd ra, 248(sp) # 8-byte Folded Spill
1514; REMAT-NEXT:    li a0, 13
1515; REMAT-NEXT:    slli a0, a0, 10
1516; REMAT-NEXT:    add a0, a1, a0
1517; REMAT-NEXT:    sd a0, 240(sp) # 8-byte Folded Spill
1518; REMAT-NEXT:    li a0, 27
1519; REMAT-NEXT:    slli a0, a0, 9
1520; REMAT-NEXT:    add a0, a1, a0
1521; REMAT-NEXT:    sd a0, 232(sp) # 8-byte Folded Spill
1522; REMAT-NEXT:    add a5, a1, a5
1523; REMAT-NEXT:    sd a5, 224(sp) # 8-byte Folded Spill
1524; REMAT-NEXT:    li a0, 29
1525; REMAT-NEXT:    slli a0, a0, 9
1526; REMAT-NEXT:    add a0, a1, a0
1527; REMAT-NEXT:    sd a0, 216(sp) # 8-byte Folded Spill
1528; REMAT-NEXT:    li a0, 15
1529; REMAT-NEXT:    slli a0, a0, 10
1530; REMAT-NEXT:    add a0, a1, a0
1531; REMAT-NEXT:    sd a0, 208(sp) # 8-byte Folded Spill
1532; REMAT-NEXT:    li a0, 31
1533; REMAT-NEXT:    slli a0, a0, 9
1534; REMAT-NEXT:    add a0, a1, a0
1535; REMAT-NEXT:    sd a0, 200(sp) # 8-byte Folded Spill
1536; REMAT-NEXT:    lui a0, 4
1537; REMAT-NEXT:    add a0, a1, a0
1538; REMAT-NEXT:    sd a0, 192(sp) # 8-byte Folded Spill
1539; REMAT-NEXT:    lui a0, 4
1540; REMAT-NEXT:    addiw a0, a0, 512
1541; REMAT-NEXT:    add a0, a1, a0
1542; REMAT-NEXT:    sd a0, 184(sp) # 8-byte Folded Spill
1543; REMAT-NEXT:    li a0, 17
1544; REMAT-NEXT:    slli a0, a0, 10
1545; REMAT-NEXT:    add a0, a1, a0
1546; REMAT-NEXT:    sd a0, 176(sp) # 8-byte Folded Spill
1547; REMAT-NEXT:    lui a0, 4
1548; REMAT-NEXT:    addiw a0, a0, 1536
1549; REMAT-NEXT:    add a0, a1, a0
1550; REMAT-NEXT:    sd a0, 168(sp) # 8-byte Folded Spill
1551; REMAT-NEXT:    li a0, 9
1552; REMAT-NEXT:    slli a0, a0, 11
1553; REMAT-NEXT:    add a0, a1, a0
1554; REMAT-NEXT:    sd a0, 160(sp) # 8-byte Folded Spill
1555; REMAT-NEXT:    lui a0, 5
1556; REMAT-NEXT:    addiw a0, a0, -1536
1557; REMAT-NEXT:    add a0, a1, a0
1558; REMAT-NEXT:    sd a0, 152(sp) # 8-byte Folded Spill
1559; REMAT-NEXT:    li a0, 19
1560; REMAT-NEXT:    slli a0, a0, 10
1561; REMAT-NEXT:    add a0, a1, a0
1562; REMAT-NEXT:    sd a0, 144(sp) # 8-byte Folded Spill
1563; REMAT-NEXT:    lui a0, 5
1564; REMAT-NEXT:    addiw a0, a0, -512
1565; REMAT-NEXT:    add a0, a1, a0
1566; REMAT-NEXT:    sd a0, 136(sp) # 8-byte Folded Spill
1567; REMAT-NEXT:    lui a0, 5
1568; REMAT-NEXT:    add a0, a1, a0
1569; REMAT-NEXT:    sd a0, 128(sp) # 8-byte Folded Spill
1570; REMAT-NEXT:    lui a0, 5
1571; REMAT-NEXT:    addiw a0, a0, 512
1572; REMAT-NEXT:    add a0, a1, a0
1573; REMAT-NEXT:    sd a0, 120(sp) # 8-byte Folded Spill
1574; REMAT-NEXT:    add s7, a1, s7
1575; REMAT-NEXT:    sd s7, 112(sp) # 8-byte Folded Spill
1576; REMAT-NEXT:    add s4, a1, s4
1577; REMAT-NEXT:    sd s4, 104(sp) # 8-byte Folded Spill
1578; REMAT-NEXT:    li a0, 11
1579; REMAT-NEXT:    slli a0, a0, 11
1580; REMAT-NEXT:    add a0, a1, a0
1581; REMAT-NEXT:    sd a0, 96(sp) # 8-byte Folded Spill
1582; REMAT-NEXT:    add s3, a1, s3
1583; REMAT-NEXT:    sd s3, 88(sp) # 8-byte Folded Spill
1584; REMAT-NEXT:    add s2, a1, s2
1585; REMAT-NEXT:    sd s2, 80(sp) # 8-byte Folded Spill
1586; REMAT-NEXT:    lui a0, 6
1587; REMAT-NEXT:    addiw a0, a0, -512
1588; REMAT-NEXT:    add a0, a1, a0
1589; REMAT-NEXT:    sd a0, 72(sp) # 8-byte Folded Spill
1590; REMAT-NEXT:    add s1, a1, s1
1591; REMAT-NEXT:    sd s1, 64(sp) # 8-byte Folded Spill
1592; REMAT-NEXT:    add s0, a1, s0
1593; REMAT-NEXT:    sd s0, 56(sp) # 8-byte Folded Spill
1594; REMAT-NEXT:    li a0, 25
1595; REMAT-NEXT:    slli a0, a0, 10
1596; REMAT-NEXT:    add a0, a1, a0
1597; REMAT-NEXT:    sd a0, 48(sp) # 8-byte Folded Spill
1598; REMAT-NEXT:    add t6, a1, t6
1599; REMAT-NEXT:    sd t6, 40(sp) # 8-byte Folded Spill
1600; REMAT-NEXT:    add t5, a1, t5
1601; REMAT-NEXT:    sd t5, 32(sp) # 8-byte Folded Spill
1602; REMAT-NEXT:    lui a0, 7
1603; REMAT-NEXT:    addiw a0, a0, -1536
1604; REMAT-NEXT:    add a0, a1, a0
1605; REMAT-NEXT:    sd a0, 24(sp) # 8-byte Folded Spill
1606; REMAT-NEXT:    add t4, a1, t4
1607; REMAT-NEXT:    sd t4, 16(sp) # 8-byte Folded Spill
1608; REMAT-NEXT:    lui ra, 7
1609; REMAT-NEXT:    addiw ra, ra, -512
1610; REMAT-NEXT:    add ra, a1, ra
1611; REMAT-NEXT:    add s11, a1, t3
1612; REMAT-NEXT:    add s10, a1, t2
1613; REMAT-NEXT:    add s9, a1, t1
1614; REMAT-NEXT:    add s8, a1, t0
1615; REMAT-NEXT:    add s7, a1, a7
1616; REMAT-NEXT:    add s6, a1, a6
1617; REMAT-NEXT:    add s5, a1, a4
1618; REMAT-NEXT:    add s4, a1, a3
1619; REMAT-NEXT:    add s3, a1, a2
1620; REMAT-NEXT:    lui s2, 8
1621; REMAT-NEXT:    addiw s2, s2, 512
1622; REMAT-NEXT:    add s2, a1, s2
1623; REMAT-NEXT:    lui s1, 8
1624; REMAT-NEXT:    addiw s1, s1, 1024
1625; REMAT-NEXT:    add s1, a1, s1
1626; REMAT-NEXT:    lui s0, 8
1627; REMAT-NEXT:    addiw s0, s0, 1536
1628; REMAT-NEXT:    add s0, a1, s0
1629; REMAT-NEXT:    li t6, 17
1630; REMAT-NEXT:    slli t6, t6, 11
1631; REMAT-NEXT:    add t6, a1, t6
1632; REMAT-NEXT:    lui t5, 9
1633; REMAT-NEXT:    addiw t5, t5, -1536
1634; REMAT-NEXT:    add t5, a1, t5
1635; REMAT-NEXT:    lui t4, 9
1636; REMAT-NEXT:    addiw t4, t4, -1024
1637; REMAT-NEXT:    add t4, a1, t4
1638; REMAT-NEXT:    lui t3, 9
1639; REMAT-NEXT:    addiw t3, t3, -512
1640; REMAT-NEXT:    add t3, a1, t3
1641; REMAT-NEXT:    lui t2, 9
1642; REMAT-NEXT:    add t2, a1, t2
1643; REMAT-NEXT:    lui t1, 9
1644; REMAT-NEXT:    addiw t1, t1, 512
1645; REMAT-NEXT:    add t1, a1, t1
1646; REMAT-NEXT:    lui t0, 9
1647; REMAT-NEXT:    addiw t0, t0, 1024
1648; REMAT-NEXT:    add t0, a1, t0
1649; REMAT-NEXT:    lui a7, 9
1650; REMAT-NEXT:    addiw a7, a7, 1536
1651; REMAT-NEXT:    add a7, a1, a7
1652; REMAT-NEXT:    li a6, 19
1653; REMAT-NEXT:    slli a6, a6, 11
1654; REMAT-NEXT:    add a6, a1, a6
1655; REMAT-NEXT:    lui a5, 10
1656; REMAT-NEXT:    addiw a5, a5, -1536
1657; REMAT-NEXT:    add a5, a1, a5
1658; REMAT-NEXT:    lui a4, 10
1659; REMAT-NEXT:    addiw a4, a4, -1024
1660; REMAT-NEXT:    add a4, a1, a4
1661; REMAT-NEXT:    lui a3, 10
1662; REMAT-NEXT:    addiw a3, a3, -512
1663; REMAT-NEXT:    add a3, a1, a3
1664; REMAT-NEXT:    lui a2, 10
1665; REMAT-NEXT:    add a2, a1, a2
1666; REMAT-NEXT:    lui a0, 10
1667; REMAT-NEXT:    addiw a0, a0, 512
1668; REMAT-NEXT:    add a0, a1, a0
1669; REMAT-NEXT:    addi a1, a1, 1536
1670; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1671; REMAT-NEXT:    vse32.v v8, (a1)
1672; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1673; REMAT-NEXT:    ld a1, 416(sp) # 8-byte Folded Reload
1674; REMAT-NEXT:    vse32.v v8, (a1)
1675; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1676; REMAT-NEXT:    ld a1, 408(sp) # 8-byte Folded Reload
1677; REMAT-NEXT:    vse32.v v8, (a1)
1678; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1679; REMAT-NEXT:    ld a1, 400(sp) # 8-byte Folded Reload
1680; REMAT-NEXT:    vse32.v v8, (a1)
1681; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1682; REMAT-NEXT:    ld a1, 392(sp) # 8-byte Folded Reload
1683; REMAT-NEXT:    vse32.v v8, (a1)
1684; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1685; REMAT-NEXT:    ld a1, 384(sp) # 8-byte Folded Reload
1686; REMAT-NEXT:    vse32.v v8, (a1)
1687; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1688; REMAT-NEXT:    ld a1, 376(sp) # 8-byte Folded Reload
1689; REMAT-NEXT:    vse32.v v8, (a1)
1690; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1691; REMAT-NEXT:    ld a1, 368(sp) # 8-byte Folded Reload
1692; REMAT-NEXT:    vse32.v v8, (a1)
1693; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1694; REMAT-NEXT:    ld a1, 360(sp) # 8-byte Folded Reload
1695; REMAT-NEXT:    vse32.v v8, (a1)
1696; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1697; REMAT-NEXT:    ld a1, 352(sp) # 8-byte Folded Reload
1698; REMAT-NEXT:    vse32.v v8, (a1)
1699; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1700; REMAT-NEXT:    ld a1, 344(sp) # 8-byte Folded Reload
1701; REMAT-NEXT:    vse32.v v8, (a1)
1702; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1703; REMAT-NEXT:    ld a1, 336(sp) # 8-byte Folded Reload
1704; REMAT-NEXT:    vse32.v v8, (a1)
1705; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1706; REMAT-NEXT:    ld a1, 328(sp) # 8-byte Folded Reload
1707; REMAT-NEXT:    vse32.v v8, (a1)
1708; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1709; REMAT-NEXT:    ld a1, 320(sp) # 8-byte Folded Reload
1710; REMAT-NEXT:    vse32.v v8, (a1)
1711; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1712; REMAT-NEXT:    ld a1, 312(sp) # 8-byte Folded Reload
1713; REMAT-NEXT:    vse32.v v8, (a1)
1714; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1715; REMAT-NEXT:    ld a1, 304(sp) # 8-byte Folded Reload
1716; REMAT-NEXT:    vse32.v v8, (a1)
1717; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1718; REMAT-NEXT:    ld a1, 296(sp) # 8-byte Folded Reload
1719; REMAT-NEXT:    vse32.v v8, (a1)
1720; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1721; REMAT-NEXT:    ld a1, 288(sp) # 8-byte Folded Reload
1722; REMAT-NEXT:    vse32.v v8, (a1)
1723; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1724; REMAT-NEXT:    ld a1, 280(sp) # 8-byte Folded Reload
1725; REMAT-NEXT:    vse32.v v8, (a1)
1726; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1727; REMAT-NEXT:    ld a1, 272(sp) # 8-byte Folded Reload
1728; REMAT-NEXT:    vse32.v v8, (a1)
1729; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1730; REMAT-NEXT:    ld a1, 264(sp) # 8-byte Folded Reload
1731; REMAT-NEXT:    vse32.v v8, (a1)
1732; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1733; REMAT-NEXT:    ld a1, 256(sp) # 8-byte Folded Reload
1734; REMAT-NEXT:    vse32.v v8, (a1)
1735; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1736; REMAT-NEXT:    ld a1, 248(sp) # 8-byte Folded Reload
1737; REMAT-NEXT:    vse32.v v8, (a1)
1738; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1739; REMAT-NEXT:    ld a1, 240(sp) # 8-byte Folded Reload
1740; REMAT-NEXT:    vse32.v v8, (a1)
1741; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1742; REMAT-NEXT:    ld a1, 232(sp) # 8-byte Folded Reload
1743; REMAT-NEXT:    vse32.v v8, (a1)
1744; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1745; REMAT-NEXT:    ld a1, 224(sp) # 8-byte Folded Reload
1746; REMAT-NEXT:    vse32.v v8, (a1)
1747; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1748; REMAT-NEXT:    ld a1, 216(sp) # 8-byte Folded Reload
1749; REMAT-NEXT:    vse32.v v8, (a1)
1750; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1751; REMAT-NEXT:    ld a1, 208(sp) # 8-byte Folded Reload
1752; REMAT-NEXT:    vse32.v v8, (a1)
1753; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1754; REMAT-NEXT:    ld a1, 200(sp) # 8-byte Folded Reload
1755; REMAT-NEXT:    vse32.v v8, (a1)
1756; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1757; REMAT-NEXT:    ld a1, 192(sp) # 8-byte Folded Reload
1758; REMAT-NEXT:    vse32.v v8, (a1)
1759; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1760; REMAT-NEXT:    ld a1, 184(sp) # 8-byte Folded Reload
1761; REMAT-NEXT:    vse32.v v8, (a1)
1762; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1763; REMAT-NEXT:    ld a1, 176(sp) # 8-byte Folded Reload
1764; REMAT-NEXT:    vse32.v v8, (a1)
1765; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1766; REMAT-NEXT:    ld a1, 168(sp) # 8-byte Folded Reload
1767; REMAT-NEXT:    vse32.v v8, (a1)
1768; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1769; REMAT-NEXT:    ld a1, 160(sp) # 8-byte Folded Reload
1770; REMAT-NEXT:    vse32.v v8, (a1)
1771; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1772; REMAT-NEXT:    ld a1, 152(sp) # 8-byte Folded Reload
1773; REMAT-NEXT:    vse32.v v8, (a1)
1774; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1775; REMAT-NEXT:    ld a1, 144(sp) # 8-byte Folded Reload
1776; REMAT-NEXT:    vse32.v v8, (a1)
1777; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1778; REMAT-NEXT:    ld a1, 136(sp) # 8-byte Folded Reload
1779; REMAT-NEXT:    vse32.v v8, (a1)
1780; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1781; REMAT-NEXT:    ld a1, 128(sp) # 8-byte Folded Reload
1782; REMAT-NEXT:    vse32.v v8, (a1)
1783; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1784; REMAT-NEXT:    ld a1, 120(sp) # 8-byte Folded Reload
1785; REMAT-NEXT:    vse32.v v8, (a1)
1786; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1787; REMAT-NEXT:    ld a1, 112(sp) # 8-byte Folded Reload
1788; REMAT-NEXT:    vse32.v v8, (a1)
1789; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1790; REMAT-NEXT:    ld a1, 104(sp) # 8-byte Folded Reload
1791; REMAT-NEXT:    vse32.v v8, (a1)
1792; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1793; REMAT-NEXT:    ld a1, 96(sp) # 8-byte Folded Reload
1794; REMAT-NEXT:    vse32.v v8, (a1)
1795; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1796; REMAT-NEXT:    ld a1, 88(sp) # 8-byte Folded Reload
1797; REMAT-NEXT:    vse32.v v8, (a1)
1798; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1799; REMAT-NEXT:    ld a1, 80(sp) # 8-byte Folded Reload
1800; REMAT-NEXT:    vse32.v v8, (a1)
1801; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1802; REMAT-NEXT:    ld a1, 72(sp) # 8-byte Folded Reload
1803; REMAT-NEXT:    vse32.v v8, (a1)
1804; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1805; REMAT-NEXT:    ld a1, 64(sp) # 8-byte Folded Reload
1806; REMAT-NEXT:    vse32.v v8, (a1)
1807; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1808; REMAT-NEXT:    ld a1, 56(sp) # 8-byte Folded Reload
1809; REMAT-NEXT:    vse32.v v8, (a1)
1810; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1811; REMAT-NEXT:    ld a1, 48(sp) # 8-byte Folded Reload
1812; REMAT-NEXT:    vse32.v v8, (a1)
1813; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1814; REMAT-NEXT:    ld a1, 40(sp) # 8-byte Folded Reload
1815; REMAT-NEXT:    vse32.v v8, (a1)
1816; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1817; REMAT-NEXT:    ld a1, 32(sp) # 8-byte Folded Reload
1818; REMAT-NEXT:    vse32.v v8, (a1)
1819; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1820; REMAT-NEXT:    ld a1, 24(sp) # 8-byte Folded Reload
1821; REMAT-NEXT:    vse32.v v8, (a1)
1822; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1823; REMAT-NEXT:    ld a1, 16(sp) # 8-byte Folded Reload
1824; REMAT-NEXT:    vse32.v v8, (a1)
1825; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1826; REMAT-NEXT:    vse32.v v8, (ra)
1827; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1828; REMAT-NEXT:    vse32.v v8, (s11)
1829; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1830; REMAT-NEXT:    vse32.v v8, (s10)
1831; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1832; REMAT-NEXT:    vse32.v v8, (s9)
1833; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1834; REMAT-NEXT:    vse32.v v8, (s8)
1835; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1836; REMAT-NEXT:    vse32.v v8, (s7)
1837; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1838; REMAT-NEXT:    vse32.v v8, (s6)
1839; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1840; REMAT-NEXT:    vse32.v v8, (s5)
1841; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1842; REMAT-NEXT:    vse32.v v8, (s4)
1843; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1844; REMAT-NEXT:    vse32.v v8, (s3)
1845; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1846; REMAT-NEXT:    vse32.v v8, (s2)
1847; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1848; REMAT-NEXT:    vse32.v v8, (s1)
1849; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1850; REMAT-NEXT:    vse32.v v8, (s0)
1851; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1852; REMAT-NEXT:    vse32.v v8, (t6)
1853; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1854; REMAT-NEXT:    vse32.v v8, (t5)
1855; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1856; REMAT-NEXT:    vse32.v v8, (t4)
1857; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1858; REMAT-NEXT:    vse32.v v8, (t3)
1859; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1860; REMAT-NEXT:    vse32.v v8, (t2)
1861; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1862; REMAT-NEXT:    vse32.v v8, (t1)
1863; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1864; REMAT-NEXT:    vse32.v v8, (t0)
1865; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1866; REMAT-NEXT:    vse32.v v8, (a7)
1867; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1868; REMAT-NEXT:    vse32.v v8, (a6)
1869; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1870; REMAT-NEXT:    vse32.v v8, (a5)
1871; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1872; REMAT-NEXT:    vse32.v v8, (a4)
1873; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1874; REMAT-NEXT:    vse32.v v8, (a3)
1875; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1876; REMAT-NEXT:    vse32.v v8, (a2)
1877; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1878; REMAT-NEXT:    vse32.v v8, (a0)
1879; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1880; REMAT-NEXT:    sf.vc.v.i 2, 0, v8, 0
1881; REMAT-NEXT:    csrr a0, vlenb
1882; REMAT-NEXT:    li a1, 18
1883; REMAT-NEXT:    mul a0, a0, a1
1884; REMAT-NEXT:    add sp, sp, a0
1885; REMAT-NEXT:    .cfi_def_cfa sp, 544
1886; REMAT-NEXT:    ld ra, 536(sp) # 8-byte Folded Reload
1887; REMAT-NEXT:    ld s0, 528(sp) # 8-byte Folded Reload
1888; REMAT-NEXT:    ld s1, 520(sp) # 8-byte Folded Reload
1889; REMAT-NEXT:    ld s2, 512(sp) # 8-byte Folded Reload
1890; REMAT-NEXT:    ld s3, 504(sp) # 8-byte Folded Reload
1891; REMAT-NEXT:    ld s4, 496(sp) # 8-byte Folded Reload
1892; REMAT-NEXT:    ld s5, 488(sp) # 8-byte Folded Reload
1893; REMAT-NEXT:    ld s6, 480(sp) # 8-byte Folded Reload
1894; REMAT-NEXT:    ld s7, 472(sp) # 8-byte Folded Reload
1895; REMAT-NEXT:    ld s8, 464(sp) # 8-byte Folded Reload
1896; REMAT-NEXT:    ld s9, 456(sp) # 8-byte Folded Reload
1897; REMAT-NEXT:    ld s10, 448(sp) # 8-byte Folded Reload
1898; REMAT-NEXT:    ld s11, 440(sp) # 8-byte Folded Reload
1899; REMAT-NEXT:    .cfi_restore ra
1900; REMAT-NEXT:    .cfi_restore s0
1901; REMAT-NEXT:    .cfi_restore s1
1902; REMAT-NEXT:    .cfi_restore s2
1903; REMAT-NEXT:    .cfi_restore s3
1904; REMAT-NEXT:    .cfi_restore s4
1905; REMAT-NEXT:    .cfi_restore s5
1906; REMAT-NEXT:    .cfi_restore s6
1907; REMAT-NEXT:    .cfi_restore s7
1908; REMAT-NEXT:    .cfi_restore s8
1909; REMAT-NEXT:    .cfi_restore s9
1910; REMAT-NEXT:    .cfi_restore s10
1911; REMAT-NEXT:    .cfi_restore s11
1912; REMAT-NEXT:    addi sp, sp, 544
1913; REMAT-NEXT:    .cfi_def_cfa_offset 0
1914; REMAT-NEXT:    ret
1915  %4 = tail call i64 @llvm.riscv.vsetvli.i64(i64 32, i64 2, i64 1)
1916  %5 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %0, i64 %4)
1917  %6 = getelementptr inbounds i32, ptr %0, i64 128
1918  %7 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %6, i64 %4)
1919  %8 = getelementptr inbounds i32, ptr %0, i64 256
1920  %9 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %8, i64 %4)
1921  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %5, <vscale x 4 x i32> %7, i64 %4)
1922  %10 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %8, i64 %4)
1923  %11 = getelementptr inbounds i32, ptr %0, i64 384
1924  %12 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %11, i64 %4)
1925  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %7, <vscale x 4 x i32> %9, i64 %4)
1926  %13 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %11, i64 %4)
1927  %14 = getelementptr inbounds i32, ptr %0, i64 512
1928  %15 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %14, i64 %4)
1929  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %10, <vscale x 4 x i32> %12, i64 %4)
1930  %16 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %14, i64 %4)
1931  %17 = getelementptr inbounds i32, ptr %0, i64 640
1932  %18 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %17, i64 %4)
1933  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %13, <vscale x 4 x i32> %15, i64 %4)
1934  %19 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %17, i64 %4)
1935  %20 = getelementptr inbounds i32, ptr %0, i64 768
1936  %21 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %20, i64 %4)
1937  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %16, <vscale x 4 x i32> %18, i64 %4)
1938  %22 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %20, i64 %4)
1939  %23 = getelementptr inbounds i32, ptr %0, i64 896
1940  %24 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %23, i64 %4)
1941  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %19, <vscale x 4 x i32> %21, i64 %4)
1942  %25 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %23, i64 %4)
1943  %26 = getelementptr inbounds i32, ptr %0, i64 1024
1944  %27 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %26, i64 %4)
1945  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %22, <vscale x 4 x i32> %24, i64 %4)
1946  %28 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %26, i64 %4)
1947  %29 = getelementptr inbounds i32, ptr %0, i64 1152
1948  %30 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %29, i64 %4)
1949  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %25, <vscale x 4 x i32> %27, i64 %4)
1950  %31 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %29, i64 %4)
1951  %32 = getelementptr inbounds i32, ptr %0, i64 1280
1952  %33 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %32, i64 %4)
1953  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %28, <vscale x 4 x i32> %30, i64 %4)
1954  %34 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %32, i64 %4)
1955  %35 = getelementptr inbounds i32, ptr %0, i64 1408
1956  %36 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %35, i64 %4)
1957  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %31, <vscale x 4 x i32> %33, i64 %4)
1958  %37 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %35, i64 %4)
1959  %38 = getelementptr inbounds i32, ptr %0, i64 1536
1960  %39 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %38, i64 %4)
1961  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %34, <vscale x 4 x i32> %36, i64 %4)
1962  %40 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %38, i64 %4)
1963  %41 = getelementptr inbounds i32, ptr %0, i64 1664
1964  %42 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %41, i64 %4)
1965  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %37, <vscale x 4 x i32> %39, i64 %4)
1966  %43 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %41, i64 %4)
1967  %44 = getelementptr inbounds i32, ptr %0, i64 1792
1968  %45 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %44, i64 %4)
1969  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %40, <vscale x 4 x i32> %42, i64 %4)
1970  %46 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %44, i64 %4)
1971  %47 = getelementptr inbounds i32, ptr %0, i64 1920
1972  %48 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %47, i64 %4)
1973  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %43, <vscale x 4 x i32> %45, i64 %4)
1974  %49 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %47, i64 %4)
1975  %50 = getelementptr inbounds i32, ptr %0, i64 2048
1976  %51 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %50, i64 %4)
1977  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %46, <vscale x 4 x i32> %48, i64 %4)
1978  %52 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %50, i64 %4)
1979  %53 = getelementptr inbounds i32, ptr %0, i64 2176
1980  %54 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %53, i64 %4)
1981  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %49, <vscale x 4 x i32> %51, i64 %4)
1982  %55 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %53, i64 %4)
1983  %56 = getelementptr inbounds i32, ptr %0, i64 2304
1984  %57 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %56, i64 %4)
1985  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %52, <vscale x 4 x i32> %54, i64 %4)
1986  %58 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %56, i64 %4)
1987  %59 = getelementptr inbounds i32, ptr %0, i64 2432
1988  %60 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %59, i64 %4)
1989  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %55, <vscale x 4 x i32> %57, i64 %4)
1990  %61 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %59, i64 %4)
1991  %62 = getelementptr inbounds i32, ptr %0, i64 2560
1992  %63 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %62, i64 %4)
1993  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %58, <vscale x 4 x i32> %60, i64 %4)
1994  %64 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %62, i64 %4)
1995  %65 = getelementptr inbounds i32, ptr %0, i64 2688
1996  %66 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %65, i64 %4)
1997  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %61, <vscale x 4 x i32> %63, i64 %4)
1998  %67 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %65, i64 %4)
1999  %68 = getelementptr inbounds i32, ptr %0, i64 2816
2000  %69 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %68, i64 %4)
2001  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %64, <vscale x 4 x i32> %66, i64 %4)
2002  %70 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %68, i64 %4)
2003  %71 = getelementptr inbounds i32, ptr %0, i64 2944
2004  %72 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %71, i64 %4)
2005  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %67, <vscale x 4 x i32> %69, i64 %4)
2006  %73 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %71, i64 %4)
2007  %74 = getelementptr inbounds i32, ptr %0, i64 3072
2008  %75 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %74, i64 %4)
2009  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %70, <vscale x 4 x i32> %72, i64 %4)
2010  %76 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %74, i64 %4)
2011  %77 = getelementptr inbounds i32, ptr %0, i64 3200
2012  %78 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %77, i64 %4)
2013  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %73, <vscale x 4 x i32> %75, i64 %4)
2014  %79 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %77, i64 %4)
2015  %80 = getelementptr inbounds i32, ptr %0, i64 3328
2016  %81 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %80, i64 %4)
2017  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %76, <vscale x 4 x i32> %78, i64 %4)
2018  %82 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %80, i64 %4)
2019  %83 = getelementptr inbounds i32, ptr %0, i64 3456
2020  %84 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %83, i64 %4)
2021  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %79, <vscale x 4 x i32> %81, i64 %4)
2022  %85 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %83, i64 %4)
2023  %86 = getelementptr inbounds i32, ptr %0, i64 3584
2024  %87 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %86, i64 %4)
2025  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %82, <vscale x 4 x i32> %84, i64 %4)
2026  %88 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %86, i64 %4)
2027  %89 = getelementptr inbounds i32, ptr %0, i64 3712
2028  %90 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %89, i64 %4)
2029  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %85, <vscale x 4 x i32> %87, i64 %4)
2030  %91 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %89, i64 %4)
2031  %92 = getelementptr inbounds i32, ptr %0, i64 3840
2032  %93 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %92, i64 %4)
2033  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %88, <vscale x 4 x i32> %90, i64 %4)
2034  %94 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %92, i64 %4)
2035  %95 = getelementptr inbounds i32, ptr %0, i64 3968
2036  %96 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %95, i64 %4)
2037  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %91, <vscale x 4 x i32> %93, i64 %4)
2038  %97 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %95, i64 %4)
2039  %98 = getelementptr inbounds i32, ptr %0, i64 4096
2040  %99 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %98, i64 %4)
2041  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %94, <vscale x 4 x i32> %96, i64 %4)
2042  %100 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %98, i64 %4)
2043  %101 = getelementptr inbounds i32, ptr %0, i64 4224
2044  %102 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %101, i64 %4)
2045  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %97, <vscale x 4 x i32> %99, i64 %4)
2046  %103 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %101, i64 %4)
2047  %104 = getelementptr inbounds i32, ptr %0, i64 4352
2048  %105 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %104, i64 %4)
2049  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %100, <vscale x 4 x i32> %102, i64 %4)
2050  %106 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %104, i64 %4)
2051  %107 = getelementptr inbounds i32, ptr %0, i64 4480
2052  %108 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %107, i64 %4)
2053  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %103, <vscale x 4 x i32> %105, i64 %4)
2054  %109 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %107, i64 %4)
2055  %110 = getelementptr inbounds i32, ptr %0, i64 4608
2056  %111 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %110, i64 %4)
2057  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %106, <vscale x 4 x i32> %108, i64 %4)
2058  %112 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %110, i64 %4)
2059  %113 = getelementptr inbounds i32, ptr %0, i64 4736
2060  %114 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %113, i64 %4)
2061  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %109, <vscale x 4 x i32> %111, i64 %4)
2062  %115 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %113, i64 %4)
2063  %116 = getelementptr inbounds i32, ptr %0, i64 4864
2064  %117 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %116, i64 %4)
2065  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %112, <vscale x 4 x i32> %114, i64 %4)
2066  %118 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %116, i64 %4)
2067  %119 = getelementptr inbounds i32, ptr %0, i64 4992
2068  %120 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %119, i64 %4)
2069  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %115, <vscale x 4 x i32> %117, i64 %4)
2070  %121 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %119, i64 %4)
2071  %122 = getelementptr inbounds i32, ptr %0, i64 5120
2072  %123 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %122, i64 %4)
2073  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %118, <vscale x 4 x i32> %120, i64 %4)
2074  %124 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %122, i64 %4)
2075  %125 = getelementptr inbounds i32, ptr %0, i64 5248
2076  %126 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %125, i64 %4)
2077  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %121, <vscale x 4 x i32> %123, i64 %4)
2078  %127 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %125, i64 %4)
2079  %128 = getelementptr inbounds i32, ptr %0, i64 5376
2080  %129 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %128, i64 %4)
2081  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %124, <vscale x 4 x i32> %126, i64 %4)
2082  %130 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %128, i64 %4)
2083  %131 = getelementptr inbounds i32, ptr %0, i64 5504
2084  %132 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %131, i64 %4)
2085  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %127, <vscale x 4 x i32> %129, i64 %4)
2086  %133 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %131, i64 %4)
2087  %134 = getelementptr inbounds i32, ptr %0, i64 5632
2088  %135 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %134, i64 %4)
2089  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %130, <vscale x 4 x i32> %132, i64 %4)
2090  %136 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %134, i64 %4)
2091  %137 = getelementptr inbounds i32, ptr %0, i64 5760
2092  %138 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %137, i64 %4)
2093  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %133, <vscale x 4 x i32> %135, i64 %4)
2094  %139 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %137, i64 %4)
2095  %140 = getelementptr inbounds i32, ptr %0, i64 5888
2096  %141 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %140, i64 %4)
2097  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %136, <vscale x 4 x i32> %138, i64 %4)
2098  %142 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %140, i64 %4)
2099  %143 = getelementptr inbounds i32, ptr %0, i64 6016
2100  %144 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %143, i64 %4)
2101  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %139, <vscale x 4 x i32> %141, i64 %4)
2102  %145 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %143, i64 %4)
2103  %146 = getelementptr inbounds i32, ptr %0, i64 6144
2104  %147 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %146, i64 %4)
2105  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %142, <vscale x 4 x i32> %144, i64 %4)
2106  %148 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %146, i64 %4)
2107  %149 = getelementptr inbounds i32, ptr %0, i64 6272
2108  %150 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %149, i64 %4)
2109  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %145, <vscale x 4 x i32> %147, i64 %4)
2110  %151 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %149, i64 %4)
2111  %152 = getelementptr inbounds i32, ptr %0, i64 6400
2112  %153 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %152, i64 %4)
2113  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %148, <vscale x 4 x i32> %150, i64 %4)
2114  %154 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %152, i64 %4)
2115  %155 = getelementptr inbounds i32, ptr %0, i64 6528
2116  %156 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %155, i64 %4)
2117  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %151, <vscale x 4 x i32> %153, i64 %4)
2118  %157 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %155, i64 %4)
2119  %158 = getelementptr inbounds i32, ptr %0, i64 6656
2120  %159 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %158, i64 %4)
2121  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %154, <vscale x 4 x i32> %156, i64 %4)
2122  %160 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %158, i64 %4)
2123  %161 = getelementptr inbounds i32, ptr %0, i64 6784
2124  %162 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %161, i64 %4)
2125  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %157, <vscale x 4 x i32> %159, i64 %4)
2126  %163 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %161, i64 %4)
2127  %164 = getelementptr inbounds i32, ptr %0, i64 6912
2128  %165 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %164, i64 %4)
2129  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %160, <vscale x 4 x i32> %162, i64 %4)
2130  %166 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %164, i64 %4)
2131  %167 = getelementptr inbounds i32, ptr %0, i64 7040
2132  %168 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %167, i64 %4)
2133  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %163, <vscale x 4 x i32> %165, i64 %4)
2134  %169 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %167, i64 %4)
2135  %170 = getelementptr inbounds i32, ptr %0, i64 7168
2136  %171 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %170, i64 %4)
2137  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %166, <vscale x 4 x i32> %168, i64 %4)
2138  %172 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %170, i64 %4)
2139  %173 = getelementptr inbounds i32, ptr %0, i64 7296
2140  %174 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %173, i64 %4)
2141  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %169, <vscale x 4 x i32> %171, i64 %4)
2142  %175 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %173, i64 %4)
2143  %176 = getelementptr inbounds i32, ptr %0, i64 7424
2144  %177 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %176, i64 %4)
2145  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %172, <vscale x 4 x i32> %174, i64 %4)
2146  %178 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %176, i64 %4)
2147  %179 = getelementptr inbounds i32, ptr %0, i64 7552
2148  %180 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %179, i64 %4)
2149  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %175, <vscale x 4 x i32> %177, i64 %4)
2150  %181 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %179, i64 %4)
2151  %182 = getelementptr inbounds i32, ptr %0, i64 7680
2152  %183 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %182, i64 %4)
2153  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %178, <vscale x 4 x i32> %180, i64 %4)
2154  %184 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %182, i64 %4)
2155  %185 = getelementptr inbounds i32, ptr %0, i64 7808
2156  %186 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %185, i64 %4)
2157  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %181, <vscale x 4 x i32> %183, i64 %4)
2158  %187 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %185, i64 %4)
2159  %188 = getelementptr inbounds i32, ptr %0, i64 7936
2160  %189 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %188, i64 %4)
2161  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %184, <vscale x 4 x i32> %186, i64 %4)
2162  %190 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %188, i64 %4)
2163  %191 = getelementptr inbounds i32, ptr %0, i64 8064
2164  %192 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %191, i64 %4)
2165  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %187, <vscale x 4 x i32> %189, i64 %4)
2166  %193 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %191, i64 %4)
2167  %194 = getelementptr inbounds i32, ptr %0, i64 8192
2168  %195 = tail call <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32> poison, ptr %194, i64 %4)
2169  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %190, <vscale x 4 x i32> %192, i64 %4)
2170  tail call void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64 3, i64 0, <vscale x 4 x i32> %193, <vscale x 4 x i32> %195, i64 %4)
2171  %196 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2172  %197 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2173  %198 = getelementptr inbounds i32, ptr %1, i64 256
2174  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %196, ptr %198, i64 %4)
2175  %199 = getelementptr inbounds i32, ptr %1, i64 384
2176  %200 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2177  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %197, ptr %199, i64 %4)
2178  %201 = getelementptr inbounds i32, ptr %1, i64 512
2179  %202 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2180  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %200, ptr %201, i64 %4)
2181  %203 = getelementptr inbounds i32, ptr %1, i64 640
2182  %204 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2183  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %202, ptr %203, i64 %4)
2184  %205 = getelementptr inbounds i32, ptr %1, i64 768
2185  %206 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2186  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %204, ptr %205, i64 %4)
2187  %207 = getelementptr inbounds i32, ptr %1, i64 896
2188  %208 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2189  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %206, ptr %207, i64 %4)
2190  %209 = getelementptr inbounds i32, ptr %1, i64 1024
2191  %210 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2192  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %208, ptr %209, i64 %4)
2193  %211 = getelementptr inbounds i32, ptr %1, i64 1152
2194  %212 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2195  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %210, ptr %211, i64 %4)
2196  %213 = getelementptr inbounds i32, ptr %1, i64 1280
2197  %214 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2198  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %212, ptr %213, i64 %4)
2199  %215 = getelementptr inbounds i32, ptr %1, i64 1408
2200  %216 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2201  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %214, ptr %215, i64 %4)
2202  %217 = getelementptr inbounds i32, ptr %1, i64 1536
2203  %218 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2204  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %216, ptr %217, i64 %4)
2205  %219 = getelementptr inbounds i32, ptr %1, i64 1664
2206  %220 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2207  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %218, ptr %219, i64 %4)
2208  %221 = getelementptr inbounds i32, ptr %1, i64 1792
2209  %222 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2210  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %220, ptr %221, i64 %4)
2211  %223 = getelementptr inbounds i32, ptr %1, i64 1920
2212  %224 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2213  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %222, ptr %223, i64 %4)
2214  %225 = getelementptr inbounds i32, ptr %1, i64 2048
2215  %226 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2216  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %224, ptr %225, i64 %4)
2217  %227 = getelementptr inbounds i32, ptr %1, i64 2176
2218  %228 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2219  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %226, ptr %227, i64 %4)
2220  %229 = getelementptr inbounds i32, ptr %1, i64 2304
2221  %230 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2222  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %228, ptr %229, i64 %4)
2223  %231 = getelementptr inbounds i32, ptr %1, i64 2432
2224  %232 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2225  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %230, ptr %231, i64 %4)
2226  %233 = getelementptr inbounds i32, ptr %1, i64 2560
2227  %234 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2228  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %232, ptr %233, i64 %4)
2229  %235 = getelementptr inbounds i32, ptr %1, i64 2688
2230  %236 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2231  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %234, ptr %235, i64 %4)
2232  %237 = getelementptr inbounds i32, ptr %1, i64 2816
2233  %238 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2234  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %236, ptr %237, i64 %4)
2235  %239 = getelementptr inbounds i32, ptr %1, i64 2944
2236  %240 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2237  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %238, ptr %239, i64 %4)
2238  %241 = getelementptr inbounds i32, ptr %1, i64 3072
2239  %242 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2240  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %240, ptr %241, i64 %4)
2241  %243 = getelementptr inbounds i32, ptr %1, i64 3200
2242  %244 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2243  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %242, ptr %243, i64 %4)
2244  %245 = getelementptr inbounds i32, ptr %1, i64 3328
2245  %246 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2246  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %244, ptr %245, i64 %4)
2247  %247 = getelementptr inbounds i32, ptr %1, i64 3456
2248  %248 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2249  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %246, ptr %247, i64 %4)
2250  %249 = getelementptr inbounds i32, ptr %1, i64 3584
2251  %250 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2252  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %248, ptr %249, i64 %4)
2253  %251 = getelementptr inbounds i32, ptr %1, i64 3712
2254  %252 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2255  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %250, ptr %251, i64 %4)
2256  %253 = getelementptr inbounds i32, ptr %1, i64 3840
2257  %254 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2258  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %252, ptr %253, i64 %4)
2259  %255 = getelementptr inbounds i32, ptr %1, i64 3968
2260  %256 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2261  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %254, ptr %255, i64 %4)
2262  %257 = getelementptr inbounds i32, ptr %1, i64 4096
2263  %258 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2264  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %256, ptr %257, i64 %4)
2265  %259 = getelementptr inbounds i32, ptr %1, i64 4224
2266  %260 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2267  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %258, ptr %259, i64 %4)
2268  %261 = getelementptr inbounds i32, ptr %1, i64 4352
2269  %262 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2270  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %260, ptr %261, i64 %4)
2271  %263 = getelementptr inbounds i32, ptr %1, i64 4480
2272  %264 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2273  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %262, ptr %263, i64 %4)
2274  %265 = getelementptr inbounds i32, ptr %1, i64 4608
2275  %266 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2276  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %264, ptr %265, i64 %4)
2277  %267 = getelementptr inbounds i32, ptr %1, i64 4736
2278  %268 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2279  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %266, ptr %267, i64 %4)
2280  %269 = getelementptr inbounds i32, ptr %1, i64 4864
2281  %270 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2282  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %268, ptr %269, i64 %4)
2283  %271 = getelementptr inbounds i32, ptr %1, i64 4992
2284  %272 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2285  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %270, ptr %271, i64 %4)
2286  %273 = getelementptr inbounds i32, ptr %1, i64 5120
2287  %274 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2288  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %272, ptr %273, i64 %4)
2289  %275 = getelementptr inbounds i32, ptr %1, i64 5248
2290  %276 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2291  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %274, ptr %275, i64 %4)
2292  %277 = getelementptr inbounds i32, ptr %1, i64 5376
2293  %278 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2294  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %276, ptr %277, i64 %4)
2295  %279 = getelementptr inbounds i32, ptr %1, i64 5504
2296  %280 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2297  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %278, ptr %279, i64 %4)
2298  %281 = getelementptr inbounds i32, ptr %1, i64 5632
2299  %282 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2300  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %280, ptr %281, i64 %4)
2301  %283 = getelementptr inbounds i32, ptr %1, i64 5760
2302  %284 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2303  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %282, ptr %283, i64 %4)
2304  %285 = getelementptr inbounds i32, ptr %1, i64 5888
2305  %286 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2306  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %284, ptr %285, i64 %4)
2307  %287 = getelementptr inbounds i32, ptr %1, i64 6016
2308  %288 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2309  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %286, ptr %287, i64 %4)
2310  %289 = getelementptr inbounds i32, ptr %1, i64 6144
2311  %290 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2312  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %288, ptr %289, i64 %4)
2313  %291 = getelementptr inbounds i32, ptr %1, i64 6272
2314  %292 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2315  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %290, ptr %291, i64 %4)
2316  %293 = getelementptr inbounds i32, ptr %1, i64 6400
2317  %294 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2318  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %292, ptr %293, i64 %4)
2319  %295 = getelementptr inbounds i32, ptr %1, i64 6528
2320  %296 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2321  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %294, ptr %295, i64 %4)
2322  %297 = getelementptr inbounds i32, ptr %1, i64 6656
2323  %298 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2324  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %296, ptr %297, i64 %4)
2325  %299 = getelementptr inbounds i32, ptr %1, i64 6784
2326  %300 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2327  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %298, ptr %299, i64 %4)
2328  %301 = getelementptr inbounds i32, ptr %1, i64 6912
2329  %302 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2330  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %300, ptr %301, i64 %4)
2331  %303 = getelementptr inbounds i32, ptr %1, i64 7040
2332  %304 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2333  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %302, ptr %303, i64 %4)
2334  %305 = getelementptr inbounds i32, ptr %1, i64 7168
2335  %306 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2336  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %304, ptr %305, i64 %4)
2337  %307 = getelementptr inbounds i32, ptr %1, i64 7296
2338  %308 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2339  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %306, ptr %307, i64 %4)
2340  %309 = getelementptr inbounds i32, ptr %1, i64 7424
2341  %310 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2342  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %308, ptr %309, i64 %4)
2343  %311 = getelementptr inbounds i32, ptr %1, i64 7552
2344  %312 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2345  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %310, ptr %311, i64 %4)
2346  %313 = getelementptr inbounds i32, ptr %1, i64 7680
2347  %314 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2348  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %312, ptr %313, i64 %4)
2349  %315 = getelementptr inbounds i32, ptr %1, i64 7808
2350  %316 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2351  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %314, ptr %315, i64 %4)
2352  %317 = getelementptr inbounds i32, ptr %1, i64 7936
2353  %318 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2354  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %316, ptr %317, i64 %4)
2355  %319 = getelementptr inbounds i32, ptr %1, i64 8064
2356  %320 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2357  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %318, ptr %319, i64 %4)
2358  %321 = getelementptr inbounds i32, ptr %1, i64 8192
2359  %322 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2360  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %320, ptr %321, i64 %4)
2361  %323 = getelementptr inbounds i32, ptr %1, i64 8320
2362  %324 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2363  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %322, ptr %323, i64 %4)
2364  %325 = getelementptr inbounds i32, ptr %1, i64 8448
2365  %326 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2366  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %324, ptr %325, i64 %4)
2367  %327 = getelementptr inbounds i32, ptr %1, i64 8576
2368  %328 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2369  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %326, ptr %327, i64 %4)
2370  %329 = getelementptr inbounds i32, ptr %1, i64 8704
2371  %330 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2372  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %328, ptr %329, i64 %4)
2373  %331 = getelementptr inbounds i32, ptr %1, i64 8832
2374  %332 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2375  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %330, ptr %331, i64 %4)
2376  %333 = getelementptr inbounds i32, ptr %1, i64 8960
2377  %334 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2378  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %332, ptr %333, i64 %4)
2379  %335 = getelementptr inbounds i32, ptr %1, i64 9088
2380  %336 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2381  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %334, ptr %335, i64 %4)
2382  %337 = getelementptr inbounds i32, ptr %1, i64 9216
2383  %338 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2384  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %336, ptr %337, i64 %4)
2385  %339 = getelementptr inbounds i32, ptr %1, i64 9344
2386  %340 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2387  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %338, ptr %339, i64 %4)
2388  %341 = getelementptr inbounds i32, ptr %1, i64 9472
2389  %342 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2390  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %340, ptr %341, i64 %4)
2391  %343 = getelementptr inbounds i32, ptr %1, i64 9600
2392  %344 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2393  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %342, ptr %343, i64 %4)
2394  %345 = getelementptr inbounds i32, ptr %1, i64 9728
2395  %346 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2396  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %344, ptr %345, i64 %4)
2397  %347 = getelementptr inbounds i32, ptr %1, i64 9856
2398  %348 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2399  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %346, ptr %347, i64 %4)
2400  %349 = getelementptr inbounds i32, ptr %1, i64 9984
2401  %350 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2402  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %348, ptr %349, i64 %4)
2403  %351 = getelementptr inbounds i32, ptr %1, i64 10112
2404  %352 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2405  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %350, ptr %351, i64 %4)
2406  %353 = getelementptr inbounds i32, ptr %1, i64 10240
2407  %354 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2408  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %352, ptr %353, i64 %4)
2409  %355 = getelementptr inbounds i32, ptr %1, i64 10368
2410  %356 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2411  tail call void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32> %354, ptr %355, i64 %4)
2412  %357 = tail call <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64 2, i64 0, i64 0, i64 %4)
2413  ret void
2414}
2415
2416declare i64 @llvm.riscv.vsetvli.i64(i64, i64, i64)
2417declare <vscale x 4 x i32> @llvm.riscv.vle.nxv4i32.i64(<vscale x 4 x i32>, ptr, i64)
2418declare void @llvm.riscv.sf.vc.vv.se.i64.nxv4i32.nxv4i32.i64(i64, i64, <vscale x 4 x i32>, <vscale x 4 x i32>, i64)
2419declare <vscale x 4 x i32> @llvm.riscv.sf.vc.v.i.se.nxv4i32.i64.i64.i64(i64, i64, i64, i64)
2420declare void @llvm.riscv.vse.nxv4i32.i64(<vscale x 4 x i32>, ptr, i64)
2421