@@ -54,6 +54,7 @@ struct bpf_jit {
54
54
int prologue_plt ; /* Start of prologue hotpatch PLT */
55
55
int kern_arena ; /* Pool offset of kernel arena address */
56
56
u64 user_arena ; /* User arena address */
57
+ u32 frame_off ; /* Offset of frame from %r15 */
57
58
};
58
59
59
60
#define SEEN_MEM BIT(0) /* use mem[] for temporary storage */
@@ -443,12 +444,9 @@ static void save_regs(struct bpf_jit *jit, u32 rs, u32 re)
443
444
/*
444
445
* Restore registers from "rs" (register start) to "re" (register end) on stack
445
446
*/
446
- static void restore_regs (struct bpf_jit * jit , u32 rs , u32 re , u32 stack_depth )
447
+ static void restore_regs (struct bpf_jit * jit , u32 rs , u32 re )
447
448
{
448
- u32 off = STK_OFF_R6 + (rs - 6 ) * 8 ;
449
-
450
- if (jit -> seen & SEEN_STACK )
451
- off += STK_OFF + stack_depth ;
449
+ u32 off = jit -> frame_off + STK_OFF_R6 + (rs - 6 ) * 8 ;
452
450
453
451
if (rs == re )
454
452
/* lg %rs,off(%r15) */
@@ -492,8 +490,7 @@ static int get_end(u16 seen_regs, int start)
492
490
* Save and restore clobbered registers (6-15) on stack.
493
491
* We save/restore registers in chunks with gap >= 2 registers.
494
492
*/
495
- static void save_restore_regs (struct bpf_jit * jit , int op , u32 stack_depth ,
496
- u16 extra_regs )
493
+ static void save_restore_regs (struct bpf_jit * jit , int op , u16 extra_regs )
497
494
{
498
495
u16 seen_regs = jit -> seen_regs | extra_regs ;
499
496
const int last = 15 , save_restore_size = 6 ;
@@ -516,7 +513,7 @@ static void save_restore_regs(struct bpf_jit *jit, int op, u32 stack_depth,
516
513
if (op == REGS_SAVE )
517
514
save_regs (jit , rs , re );
518
515
else
519
- restore_regs (jit , rs , re , stack_depth );
516
+ restore_regs (jit , rs , re );
520
517
re ++ ;
521
518
} while (re <= last );
522
519
}
@@ -575,8 +572,7 @@ static void bpf_jit_plt(struct bpf_plt *plt, void *ret, void *target)
575
572
* Save registers and create stack frame if necessary.
576
573
* See stack frame layout description in "bpf_jit.h"!
577
574
*/
578
- static void bpf_jit_prologue (struct bpf_jit * jit , struct bpf_prog * fp ,
579
- u32 stack_depth )
575
+ static void bpf_jit_prologue (struct bpf_jit * jit , struct bpf_prog * fp )
580
576
{
581
577
/* No-op for hotpatching */
582
578
/* brcl 0,prologue_plt */
@@ -609,7 +605,7 @@ static void bpf_jit_prologue(struct bpf_jit *jit, struct bpf_prog *fp,
609
605
jit -> seen_regs |= NVREGS ;
610
606
} else {
611
607
/* Save registers */
612
- save_restore_regs (jit , REGS_SAVE , stack_depth ,
608
+ save_restore_regs (jit , REGS_SAVE ,
613
609
fp -> aux -> exception_boundary ? NVREGS : 0 );
614
610
}
615
611
/* Setup literal pool */
@@ -631,8 +627,8 @@ static void bpf_jit_prologue(struct bpf_jit *jit, struct bpf_prog *fp,
631
627
EMIT4 (0xb9040000 , REG_W1 , REG_15 );
632
628
/* la %bfp,STK_160_UNUSED(%r15) (BPF frame pointer) */
633
629
EMIT4_DISP (0x41000000 , BPF_REG_FP , REG_15 , STK_160_UNUSED );
634
- /* aghi %r15,-STK_OFF */
635
- EMIT4_IMM (0xa70b0000 , REG_15 , - ( STK_OFF + stack_depth ) );
630
+ /* aghi %r15,-frame_off */
631
+ EMIT4_IMM (0xa70b0000 , REG_15 , - jit -> frame_off );
636
632
/* stg %w1,152(%r15) (backchain) */
637
633
EMIT6_DISP_LH (0xe3000000 , 0x0024 , REG_W1 , REG_0 ,
638
634
REG_15 , 152 );
@@ -669,13 +665,13 @@ static void call_r1(struct bpf_jit *jit)
669
665
/*
670
666
* Function epilogue
671
667
*/
672
- static void bpf_jit_epilogue (struct bpf_jit * jit , u32 stack_depth )
668
+ static void bpf_jit_epilogue (struct bpf_jit * jit )
673
669
{
674
670
jit -> exit_ip = jit -> prg ;
675
671
/* Load exit code: lgr %r2,%b0 */
676
672
EMIT4 (0xb9040000 , REG_2 , BPF_REG_0 );
677
673
/* Restore registers */
678
- save_restore_regs (jit , REGS_RESTORE , stack_depth , 0 );
674
+ save_restore_regs (jit , REGS_RESTORE , 0 );
679
675
EMIT_JUMP_REG (14 );
680
676
681
677
jit -> prg = ALIGN (jit -> prg , 8 );
@@ -857,7 +853,7 @@ static int sign_extend(struct bpf_jit *jit, int r, u8 size, u8 flags)
857
853
* stack space for the large switch statement.
858
854
*/
859
855
static noinline int bpf_jit_insn (struct bpf_jit * jit , struct bpf_prog * fp ,
860
- int i , bool extra_pass , u32 stack_depth )
856
+ int i , bool extra_pass )
861
857
{
862
858
struct bpf_insn * insn = & fp -> insnsi [i ];
863
859
s32 branch_oc_off = insn -> off ;
@@ -1778,9 +1774,9 @@ static noinline int bpf_jit_insn(struct bpf_jit *jit, struct bpf_prog *fp,
1778
1774
* Note 2: We assume that the verifier does not let us call the
1779
1775
* main program, which clears the tail call counter on entry.
1780
1776
*/
1781
- /* mvc STK_OFF_TCCNT(4,%r15),N (%r15) */
1777
+ /* mvc STK_OFF_TCCNT(4,%r15),frame_off+STK_OFF_TCCNT (%r15) */
1782
1778
_EMIT6 (0xd203f000 | STK_OFF_TCCNT ,
1783
- 0xf000 | (STK_OFF_TCCNT + STK_OFF + stack_depth ));
1779
+ 0xf000 | (jit -> frame_off + STK_OFF_TCCNT ));
1784
1780
1785
1781
/* Sign-extend the kfunc arguments. */
1786
1782
if (insn -> src_reg == BPF_PSEUDO_KFUNC_CALL ) {
@@ -1831,10 +1827,7 @@ static noinline int bpf_jit_insn(struct bpf_jit *jit, struct bpf_prog *fp,
1831
1827
* goto out;
1832
1828
*/
1833
1829
1834
- if (jit -> seen & SEEN_STACK )
1835
- off = STK_OFF_TCCNT + STK_OFF + stack_depth ;
1836
- else
1837
- off = STK_OFF_TCCNT ;
1830
+ off = jit -> frame_off + STK_OFF_TCCNT ;
1838
1831
/* lhi %w0,1 */
1839
1832
EMIT4_IMM (0xa7080000 , REG_W0 , 1 );
1840
1833
/* laal %w1,%w0,off(%r15) */
@@ -1864,7 +1857,7 @@ static noinline int bpf_jit_insn(struct bpf_jit *jit, struct bpf_prog *fp,
1864
1857
/*
1865
1858
* Restore registers before calling function
1866
1859
*/
1867
- save_restore_regs (jit , REGS_RESTORE , stack_depth , 0 );
1860
+ save_restore_regs (jit , REGS_RESTORE , 0 );
1868
1861
1869
1862
/*
1870
1863
* goto *(prog->bpf_func + tail_call_start);
@@ -2157,7 +2150,7 @@ static int bpf_set_addr(struct bpf_jit *jit, int i)
2157
2150
* Compile eBPF program into s390x code
2158
2151
*/
2159
2152
static int bpf_jit_prog (struct bpf_jit * jit , struct bpf_prog * fp ,
2160
- bool extra_pass , u32 stack_depth )
2153
+ bool extra_pass )
2161
2154
{
2162
2155
int i , insn_count , lit32_size , lit64_size ;
2163
2156
u64 kern_arena ;
@@ -2166,24 +2159,28 @@ static int bpf_jit_prog(struct bpf_jit *jit, struct bpf_prog *fp,
2166
2159
jit -> lit64 = jit -> lit64_start ;
2167
2160
jit -> prg = 0 ;
2168
2161
jit -> excnt = 0 ;
2162
+ if (is_first_pass (jit ) || (jit -> seen & SEEN_STACK ))
2163
+ jit -> frame_off = STK_OFF + round_up (fp -> aux -> stack_depth , 8 );
2164
+ else
2165
+ jit -> frame_off = 0 ;
2169
2166
2170
2167
kern_arena = bpf_arena_get_kern_vm_start (fp -> aux -> arena );
2171
2168
if (kern_arena )
2172
2169
jit -> kern_arena = _EMIT_CONST_U64 (kern_arena );
2173
2170
jit -> user_arena = bpf_arena_get_user_vm_start (fp -> aux -> arena );
2174
2171
2175
- bpf_jit_prologue (jit , fp , stack_depth );
2172
+ bpf_jit_prologue (jit , fp );
2176
2173
if (bpf_set_addr (jit , 0 ) < 0 )
2177
2174
return -1 ;
2178
2175
for (i = 0 ; i < fp -> len ; i += insn_count ) {
2179
- insn_count = bpf_jit_insn (jit , fp , i , extra_pass , stack_depth );
2176
+ insn_count = bpf_jit_insn (jit , fp , i , extra_pass );
2180
2177
if (insn_count < 0 )
2181
2178
return -1 ;
2182
2179
/* Next instruction address */
2183
2180
if (bpf_set_addr (jit , i + insn_count ) < 0 )
2184
2181
return -1 ;
2185
2182
}
2186
- bpf_jit_epilogue (jit , stack_depth );
2183
+ bpf_jit_epilogue (jit );
2187
2184
2188
2185
lit32_size = jit -> lit32 - jit -> lit32_start ;
2189
2186
lit64_size = jit -> lit64 - jit -> lit64_start ;
@@ -2259,7 +2256,6 @@ static struct bpf_binary_header *bpf_jit_alloc(struct bpf_jit *jit,
2259
2256
*/
2260
2257
struct bpf_prog * bpf_int_jit_compile (struct bpf_prog * fp )
2261
2258
{
2262
- u32 stack_depth = round_up (fp -> aux -> stack_depth , 8 );
2263
2259
struct bpf_prog * tmp , * orig_fp = fp ;
2264
2260
struct bpf_binary_header * header ;
2265
2261
struct s390_jit_data * jit_data ;
@@ -2312,7 +2308,7 @@ struct bpf_prog *bpf_int_jit_compile(struct bpf_prog *fp)
2312
2308
* - 3: Calculate program size and addrs array
2313
2309
*/
2314
2310
for (pass = 1 ; pass <= 3 ; pass ++ ) {
2315
- if (bpf_jit_prog (& jit , fp , extra_pass , stack_depth )) {
2311
+ if (bpf_jit_prog (& jit , fp , extra_pass )) {
2316
2312
fp = orig_fp ;
2317
2313
goto free_addrs ;
2318
2314
}
@@ -2326,7 +2322,7 @@ struct bpf_prog *bpf_int_jit_compile(struct bpf_prog *fp)
2326
2322
goto free_addrs ;
2327
2323
}
2328
2324
skip_init_ctx :
2329
- if (bpf_jit_prog (& jit , fp , extra_pass , stack_depth )) {
2325
+ if (bpf_jit_prog (& jit , fp , extra_pass )) {
2330
2326
bpf_jit_binary_free (header );
2331
2327
fp = orig_fp ;
2332
2328
goto free_addrs ;
0 commit comments