Message ID | 20190531185705.2629959-3-jonathan.lemon@gmail.com |
---|---|
State | Changes Requested |
Delegated to: | BPF Maintainers |
Headers | show |
Series | Better handling of xskmap entries | expand |
> On May 31, 2019, at 11:57 AM, Jonathan Lemon <jonathan.lemon@gmail.com> wrote: > > Use the recent change to XSKMAP bpf_map_lookup_elem() to test if > there is a xsk present in the map instead of duplicating the work > with qidconf. > > Fix things so callers using XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD > bypass any internal bpf maps, so xsk_socket__{create|delete} works > properly. > > Signed-off-by: Jonathan Lemon <jonathan.lemon@gmail.com> > --- > tools/lib/bpf/xsk.c | 79 +++++++++------------------------------------ > 1 file changed, 16 insertions(+), 63 deletions(-) > > diff --git a/tools/lib/bpf/xsk.c b/tools/lib/bpf/xsk.c > index 38667b62f1fe..7ce7494b5b50 100644 > --- a/tools/lib/bpf/xsk.c > +++ b/tools/lib/bpf/xsk.c > @@ -60,10 +60,8 @@ struct xsk_socket { > struct xsk_umem *umem; > struct xsk_socket_config config; > int fd; > - int xsks_map; > int ifindex; > int prog_fd; > - int qidconf_map_fd; > int xsks_map_fd; > __u32 queue_id; > char ifname[IFNAMSIZ]; > @@ -265,15 +263,11 @@ static int xsk_load_xdp_prog(struct xsk_socket *xsk) > /* This is the C-program: > * SEC("xdp_sock") int xdp_sock_prog(struct xdp_md *ctx) > * { > - * int *qidconf, index = ctx->rx_queue_index; > + * int index = ctx->rx_queue_index; > * > * // A set entry here means that the correspnding queue_id > * // has an active AF_XDP socket bound to it. > - * qidconf = bpf_map_lookup_elem(&qidconf_map, &index); > - * if (!qidconf) > - * return XDP_ABORTED; > - * > - * if (*qidconf) > + * if (bpf_map_lookup_elem(&xsks_map, &index)) > * return bpf_redirect_map(&xsks_map, index, 0); > * > * return XDP_PASS; > @@ -286,15 +280,10 @@ static int xsk_load_xdp_prog(struct xsk_socket *xsk) > BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_1, -4), > BPF_MOV64_REG(BPF_REG_2, BPF_REG_10), > BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), > - BPF_LD_MAP_FD(BPF_REG_1, xsk->qidconf_map_fd), > + BPF_LD_MAP_FD(BPF_REG_1, xsk->xsks_map_fd), > BPF_EMIT_CALL(BPF_FUNC_map_lookup_elem), > BPF_MOV64_REG(BPF_REG_1, BPF_REG_0), > - BPF_MOV32_IMM(BPF_REG_0, 0), > - /* if r1 == 0 goto +8 */ > - BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 8), > BPF_MOV32_IMM(BPF_REG_0, 2), > - /* r1 = *(u32 *)(r1 + 0) */ > - BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_1, 0), > /* if r1 == 0 goto +5 */ > BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 5), > /* r2 = *(u32 *)(r10 - 4) */ > @@ -366,18 +355,11 @@ static int xsk_create_bpf_maps(struct xsk_socket *xsk) > if (max_queues < 0) > return max_queues; > > - fd = bpf_create_map_name(BPF_MAP_TYPE_ARRAY, "qidconf_map", > + fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map", > sizeof(int), sizeof(int), max_queues, 0); > if (fd < 0) > return fd; > - xsk->qidconf_map_fd = fd; > > - fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map", > - sizeof(int), sizeof(int), max_queues, 0); > - if (fd < 0) { > - close(xsk->qidconf_map_fd); > - return fd; > - } > xsk->xsks_map_fd = fd; > > return 0; > @@ -385,10 +367,8 @@ static int xsk_create_bpf_maps(struct xsk_socket *xsk) > > static void xsk_delete_bpf_maps(struct xsk_socket *xsk) > { > - close(xsk->qidconf_map_fd); > + bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id); > close(xsk->xsks_map_fd); > - xsk->qidconf_map_fd = -1; > - xsk->xsks_map_fd = -1; > } > > static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) > @@ -417,10 +397,9 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) > if (err) > goto out_map_ids; > > - for (i = 0; i < prog_info.nr_map_ids; i++) { > - if (xsk->qidconf_map_fd != -1 && xsk->xsks_map_fd != -1) > - break; > + xsk->xsks_map_fd = -1; > > + for (i = 0; i < prog_info.nr_map_ids; i++) { > fd = bpf_map_get_fd_by_id(map_ids[i]); > if (fd < 0) > continue; > @@ -431,11 +410,6 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) > continue; > } > > - if (!strcmp(map_info.name, "qidconf_map")) { > - xsk->qidconf_map_fd = fd; > - continue; > - } > - > if (!strcmp(map_info.name, "xsks_map")) { > xsk->xsks_map_fd = fd; > continue; > @@ -445,40 +419,18 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) > } > > err = 0; > - if (xsk->qidconf_map_fd < 0 || xsk->xsks_map_fd < 0) { > + if (xsk->xsks_map_fd == -1) > err = -ENOENT; > - xsk_delete_bpf_maps(xsk); > - } > > out_map_ids: > free(map_ids); > return err; > } > > -static void xsk_clear_bpf_maps(struct xsk_socket *xsk) > -{ > - int qid = false; > - > - bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0); > - bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id); > -} > - > static int xsk_set_bpf_maps(struct xsk_socket *xsk) > { > - int qid = true, fd = xsk->fd, err; > - > - err = bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0); > - if (err) > - goto out; > - > - err = bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, &fd, 0); > - if (err) > - goto out; > - > - return 0; > -out: > - xsk_clear_bpf_maps(xsk); > - return err; > + return bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, > + &xsk->fd, 0); > } > > static int xsk_setup_xdp_prog(struct xsk_socket *xsk) > @@ -514,6 +466,7 @@ static int xsk_setup_xdp_prog(struct xsk_socket *xsk) > > out_load: > close(xsk->prog_fd); > + xsk->prog_fd = -1; I found xsk->prog_fd confusing. Why do we need to set it here? I think we don't need to call xsk_delete_bpf_maps() in out_load path? > out_maps: > xsk_delete_bpf_maps(xsk); > return err; > @@ -643,9 +596,7 @@ int xsk_socket__create(struct xsk_socket **xsk_ptr, const char *ifname, > goto out_mmap_tx; > } > > - xsk->qidconf_map_fd = -1; > - xsk->xsks_map_fd = -1; > - > + xsk->prog_fd = -1; > if (!(xsk->config.libbpf_flags & XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) { > err = xsk_setup_xdp_prog(xsk); > if (err) > @@ -708,8 +659,10 @@ void xsk_socket__delete(struct xsk_socket *xsk) > if (!xsk) > return; > > - xsk_clear_bpf_maps(xsk); > - xsk_delete_bpf_maps(xsk); > + if (xsk->prog_fd != -1) { > + xsk_delete_bpf_maps(xsk); > + close(xsk->prog_fd); Here, we use prog_fd != -1 to gate xsk_delete_bpf_maps(), which is confusing. I looked at the code for quite sometime, but still cannot confirm it is correct. Thanks, Song > + } > > optlen = sizeof(off); > err = getsockopt(xsk->fd, SOL_XDP, XDP_MMAP_OFFSETS, &off, &optlen); > -- > 2.17.1 >
On 1 Jun 2019, at 16:05, Song Liu wrote: >> On May 31, 2019, at 11:57 AM, Jonathan Lemon >> <jonathan.lemon@gmail.com> wrote: >> >> Use the recent change to XSKMAP bpf_map_lookup_elem() to test if >> there is a xsk present in the map instead of duplicating the work >> with qidconf. >> >> Fix things so callers using XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD >> bypass any internal bpf maps, so xsk_socket__{create|delete} works >> properly. >> >> Signed-off-by: Jonathan Lemon <jonathan.lemon@gmail.com> >> --- >> tools/lib/bpf/xsk.c | 79 >> +++++++++------------------------------------ >> 1 file changed, 16 insertions(+), 63 deletions(-) >> >> diff --git a/tools/lib/bpf/xsk.c b/tools/lib/bpf/xsk.c >> index 38667b62f1fe..7ce7494b5b50 100644 >> --- a/tools/lib/bpf/xsk.c >> +++ b/tools/lib/bpf/xsk.c >> @@ -60,10 +60,8 @@ struct xsk_socket { >> struct xsk_umem *umem; >> struct xsk_socket_config config; >> int fd; >> - int xsks_map; >> int ifindex; >> int prog_fd; >> - int qidconf_map_fd; >> int xsks_map_fd; >> __u32 queue_id; >> char ifname[IFNAMSIZ]; >> @@ -265,15 +263,11 @@ static int xsk_load_xdp_prog(struct xsk_socket >> *xsk) >> /* This is the C-program: >> * SEC("xdp_sock") int xdp_sock_prog(struct xdp_md *ctx) >> * { >> - * int *qidconf, index = ctx->rx_queue_index; >> + * int index = ctx->rx_queue_index; >> * >> * // A set entry here means that the correspnding queue_id >> * // has an active AF_XDP socket bound to it. >> - * qidconf = bpf_map_lookup_elem(&qidconf_map, &index); >> - * if (!qidconf) >> - * return XDP_ABORTED; >> - * >> - * if (*qidconf) >> + * if (bpf_map_lookup_elem(&xsks_map, &index)) >> * return bpf_redirect_map(&xsks_map, index, 0); >> * >> * return XDP_PASS; >> @@ -286,15 +280,10 @@ static int xsk_load_xdp_prog(struct xsk_socket >> *xsk) >> BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_1, -4), >> BPF_MOV64_REG(BPF_REG_2, BPF_REG_10), >> BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), >> - BPF_LD_MAP_FD(BPF_REG_1, xsk->qidconf_map_fd), >> + BPF_LD_MAP_FD(BPF_REG_1, xsk->xsks_map_fd), >> BPF_EMIT_CALL(BPF_FUNC_map_lookup_elem), >> BPF_MOV64_REG(BPF_REG_1, BPF_REG_0), >> - BPF_MOV32_IMM(BPF_REG_0, 0), >> - /* if r1 == 0 goto +8 */ >> - BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 8), >> BPF_MOV32_IMM(BPF_REG_0, 2), >> - /* r1 = *(u32 *)(r1 + 0) */ >> - BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_1, 0), >> /* if r1 == 0 goto +5 */ >> BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 5), >> /* r2 = *(u32 *)(r10 - 4) */ >> @@ -366,18 +355,11 @@ static int xsk_create_bpf_maps(struct >> xsk_socket *xsk) >> if (max_queues < 0) >> return max_queues; >> >> - fd = bpf_create_map_name(BPF_MAP_TYPE_ARRAY, "qidconf_map", >> + fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map", >> sizeof(int), sizeof(int), max_queues, 0); >> if (fd < 0) >> return fd; >> - xsk->qidconf_map_fd = fd; >> >> - fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map", >> - sizeof(int), sizeof(int), max_queues, 0); >> - if (fd < 0) { >> - close(xsk->qidconf_map_fd); >> - return fd; >> - } >> xsk->xsks_map_fd = fd; >> >> return 0; >> @@ -385,10 +367,8 @@ static int xsk_create_bpf_maps(struct xsk_socket >> *xsk) >> >> static void xsk_delete_bpf_maps(struct xsk_socket *xsk) >> { >> - close(xsk->qidconf_map_fd); >> + bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id); >> close(xsk->xsks_map_fd); >> - xsk->qidconf_map_fd = -1; >> - xsk->xsks_map_fd = -1; >> } >> >> static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) >> @@ -417,10 +397,9 @@ static int xsk_lookup_bpf_maps(struct xsk_socket >> *xsk) >> if (err) >> goto out_map_ids; >> >> - for (i = 0; i < prog_info.nr_map_ids; i++) { >> - if (xsk->qidconf_map_fd != -1 && xsk->xsks_map_fd != -1) >> - break; >> + xsk->xsks_map_fd = -1; >> >> + for (i = 0; i < prog_info.nr_map_ids; i++) { >> fd = bpf_map_get_fd_by_id(map_ids[i]); >> if (fd < 0) >> continue; >> @@ -431,11 +410,6 @@ static int xsk_lookup_bpf_maps(struct xsk_socket >> *xsk) >> continue; >> } >> >> - if (!strcmp(map_info.name, "qidconf_map")) { >> - xsk->qidconf_map_fd = fd; >> - continue; >> - } >> - >> if (!strcmp(map_info.name, "xsks_map")) { >> xsk->xsks_map_fd = fd; >> continue; >> @@ -445,40 +419,18 @@ static int xsk_lookup_bpf_maps(struct >> xsk_socket *xsk) >> } >> >> err = 0; >> - if (xsk->qidconf_map_fd < 0 || xsk->xsks_map_fd < 0) { >> + if (xsk->xsks_map_fd == -1) >> err = -ENOENT; >> - xsk_delete_bpf_maps(xsk); >> - } >> >> out_map_ids: >> free(map_ids); >> return err; >> } >> >> -static void xsk_clear_bpf_maps(struct xsk_socket *xsk) >> -{ >> - int qid = false; >> - >> - bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0); >> - bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id); >> -} >> - >> static int xsk_set_bpf_maps(struct xsk_socket *xsk) >> { >> - int qid = true, fd = xsk->fd, err; >> - >> - err = bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, >> &qid, 0); >> - if (err) >> - goto out; >> - >> - err = bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, &fd, >> 0); >> - if (err) >> - goto out; >> - >> - return 0; >> -out: >> - xsk_clear_bpf_maps(xsk); >> - return err; >> + return bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, >> + &xsk->fd, 0); >> } >> >> static int xsk_setup_xdp_prog(struct xsk_socket *xsk) >> @@ -514,6 +466,7 @@ static int xsk_setup_xdp_prog(struct xsk_socket >> *xsk) >> >> out_load: >> close(xsk->prog_fd); >> + xsk->prog_fd = -1; > > I found xsk->prog_fd confusing. Why do we need to set it here? I suppose this one isn't strictly required - I set it as a guard out of habit. xsk is (currently) immediately freed by the caller, so it can be removed. The main logic is: xsk->prog_fd = -1; if (!(xsk->config.libbpf_flags & XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) { err = xsk_setup_xdp_prog(xsk); The user may pass INHIBIT_PROG_LOAD, which bypasses setting up the xdp program (and any maps associated with the program), allowing installation of a custom program. The cleanup behavior is then gated on prog_fd being -1, > > I think we don't need to call xsk_delete_bpf_maps() in out_load path? Hmm, there's two out_load paths, but only one needs the delete maps call. Let me redo the error handling so it's a bit more explicit. > >> out_maps: >> xsk_delete_bpf_maps(xsk); >> return err; >> @@ -643,9 +596,7 @@ int xsk_socket__create(struct xsk_socket >> **xsk_ptr, const char *ifname, >> goto out_mmap_tx; >> } >> >> - xsk->qidconf_map_fd = -1; >> - xsk->xsks_map_fd = -1; >> - >> + xsk->prog_fd = -1; >> if (!(xsk->config.libbpf_flags & >> XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) { >> err = xsk_setup_xdp_prog(xsk); >> if (err) >> @@ -708,8 +659,10 @@ void xsk_socket__delete(struct xsk_socket *xsk) >> if (!xsk) >> return; >> >> - xsk_clear_bpf_maps(xsk); >> - xsk_delete_bpf_maps(xsk); >> + if (xsk->prog_fd != -1) { >> + xsk_delete_bpf_maps(xsk); >> + close(xsk->prog_fd); > > Here, we use prog_fd != -1 to gate xsk_delete_bpf_maps(), which is > confusing. I looked at the code for quite sometime, but still cannot > confirm it is correct. See above reasoning - with INHIBIT_PROG_LOAD, there is no library-provided program or maps, so cleanup actions are skipped.
> On Jun 1, 2019, at 9:18 PM, Jonathan Lemon <jonathan.lemon@gmail.com> wrote: > > > > On 1 Jun 2019, at 16:05, Song Liu wrote: > >>> On May 31, 2019, at 11:57 AM, Jonathan Lemon <jonathan.lemon@gmail.com> wrote: >>> >>> Use the recent change to XSKMAP bpf_map_lookup_elem() to test if >>> there is a xsk present in the map instead of duplicating the work >>> with qidconf. >>> >>> Fix things so callers using XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD >>> bypass any internal bpf maps, so xsk_socket__{create|delete} works >>> properly. >>> >>> Signed-off-by: Jonathan Lemon <jonathan.lemon@gmail.com> >>> --- >>> tools/lib/bpf/xsk.c | 79 +++++++++------------------------------------ >>> 1 file changed, 16 insertions(+), 63 deletions(-) >>> >>> diff --git a/tools/lib/bpf/xsk.c b/tools/lib/bpf/xsk.c >>> index 38667b62f1fe..7ce7494b5b50 100644 >>> --- a/tools/lib/bpf/xsk.c >>> +++ b/tools/lib/bpf/xsk.c >>> @@ -60,10 +60,8 @@ struct xsk_socket { >>> struct xsk_umem *umem; >>> struct xsk_socket_config config; >>> int fd; >>> - int xsks_map; >>> int ifindex; >>> int prog_fd; >>> - int qidconf_map_fd; >>> int xsks_map_fd; >>> __u32 queue_id; >>> char ifname[IFNAMSIZ]; >>> @@ -265,15 +263,11 @@ static int xsk_load_xdp_prog(struct xsk_socket *xsk) >>> /* This is the C-program: >>> * SEC("xdp_sock") int xdp_sock_prog(struct xdp_md *ctx) >>> * { >>> - * int *qidconf, index = ctx->rx_queue_index; >>> + * int index = ctx->rx_queue_index; >>> * >>> * // A set entry here means that the correspnding queue_id >>> * // has an active AF_XDP socket bound to it. >>> - * qidconf = bpf_map_lookup_elem(&qidconf_map, &index); >>> - * if (!qidconf) >>> - * return XDP_ABORTED; >>> - * >>> - * if (*qidconf) >>> + * if (bpf_map_lookup_elem(&xsks_map, &index)) >>> * return bpf_redirect_map(&xsks_map, index, 0); >>> * >>> * return XDP_PASS; >>> @@ -286,15 +280,10 @@ static int xsk_load_xdp_prog(struct xsk_socket *xsk) >>> BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_1, -4), >>> BPF_MOV64_REG(BPF_REG_2, BPF_REG_10), >>> BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), >>> - BPF_LD_MAP_FD(BPF_REG_1, xsk->qidconf_map_fd), >>> + BPF_LD_MAP_FD(BPF_REG_1, xsk->xsks_map_fd), >>> BPF_EMIT_CALL(BPF_FUNC_map_lookup_elem), >>> BPF_MOV64_REG(BPF_REG_1, BPF_REG_0), >>> - BPF_MOV32_IMM(BPF_REG_0, 0), >>> - /* if r1 == 0 goto +8 */ >>> - BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 8), >>> BPF_MOV32_IMM(BPF_REG_0, 2), >>> - /* r1 = *(u32 *)(r1 + 0) */ >>> - BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_1, 0), >>> /* if r1 == 0 goto +5 */ >>> BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 5), >>> /* r2 = *(u32 *)(r10 - 4) */ >>> @@ -366,18 +355,11 @@ static int xsk_create_bpf_maps(struct xsk_socket *xsk) >>> if (max_queues < 0) >>> return max_queues; >>> >>> - fd = bpf_create_map_name(BPF_MAP_TYPE_ARRAY, "qidconf_map", >>> + fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map", >>> sizeof(int), sizeof(int), max_queues, 0); >>> if (fd < 0) >>> return fd; >>> - xsk->qidconf_map_fd = fd; >>> >>> - fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map", >>> - sizeof(int), sizeof(int), max_queues, 0); >>> - if (fd < 0) { >>> - close(xsk->qidconf_map_fd); >>> - return fd; >>> - } >>> xsk->xsks_map_fd = fd; >>> >>> return 0; >>> @@ -385,10 +367,8 @@ static int xsk_create_bpf_maps(struct xsk_socket *xsk) >>> >>> static void xsk_delete_bpf_maps(struct xsk_socket *xsk) >>> { >>> - close(xsk->qidconf_map_fd); >>> + bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id); >>> close(xsk->xsks_map_fd); >>> - xsk->qidconf_map_fd = -1; >>> - xsk->xsks_map_fd = -1; >>> } >>> >>> static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) >>> @@ -417,10 +397,9 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) >>> if (err) >>> goto out_map_ids; >>> >>> - for (i = 0; i < prog_info.nr_map_ids; i++) { >>> - if (xsk->qidconf_map_fd != -1 && xsk->xsks_map_fd != -1) >>> - break; >>> + xsk->xsks_map_fd = -1; >>> >>> + for (i = 0; i < prog_info.nr_map_ids; i++) { >>> fd = bpf_map_get_fd_by_id(map_ids[i]); >>> if (fd < 0) >>> continue; >>> @@ -431,11 +410,6 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) >>> continue; >>> } >>> >>> - if (!strcmp(map_info.name, "qidconf_map")) { >>> - xsk->qidconf_map_fd = fd; >>> - continue; >>> - } >>> - >>> if (!strcmp(map_info.name, "xsks_map")) { >>> xsk->xsks_map_fd = fd; >>> continue; >>> @@ -445,40 +419,18 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) >>> } >>> >>> err = 0; >>> - if (xsk->qidconf_map_fd < 0 || xsk->xsks_map_fd < 0) { >>> + if (xsk->xsks_map_fd == -1) >>> err = -ENOENT; >>> - xsk_delete_bpf_maps(xsk); >>> - } >>> >>> out_map_ids: >>> free(map_ids); >>> return err; >>> } >>> >>> -static void xsk_clear_bpf_maps(struct xsk_socket *xsk) >>> -{ >>> - int qid = false; >>> - >>> - bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0); >>> - bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id); >>> -} >>> - >>> static int xsk_set_bpf_maps(struct xsk_socket *xsk) >>> { >>> - int qid = true, fd = xsk->fd, err; >>> - >>> - err = bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0); >>> - if (err) >>> - goto out; >>> - >>> - err = bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, &fd, 0); >>> - if (err) >>> - goto out; >>> - >>> - return 0; >>> -out: >>> - xsk_clear_bpf_maps(xsk); >>> - return err; >>> + return bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, >>> + &xsk->fd, 0); >>> } >>> >>> static int xsk_setup_xdp_prog(struct xsk_socket *xsk) >>> @@ -514,6 +466,7 @@ static int xsk_setup_xdp_prog(struct xsk_socket *xsk) >>> >>> out_load: >>> close(xsk->prog_fd); >>> + xsk->prog_fd = -1; >> >> I found xsk->prog_fd confusing. Why do we need to set it here? > > I suppose this one isn't strictly required - I set it as a guard out of habit. > xsk is (currently) immediately freed by the caller, so it can be removed. > > > The main logic is: > > xsk->prog_fd = -1; > if (!(xsk->config.libbpf_flags & XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) { > err = xsk_setup_xdp_prog(xsk); > > The user may pass INHIBIT_PROG_LOAD, which bypasses setting up the xdp program > (and any maps associated with the program), allowing installation of a custom > program. The cleanup behavior is then gated on prog_fd being -1, > >> >> I think we don't need to call xsk_delete_bpf_maps() in out_load path? > > Hmm, there's two out_load paths, but only one needs the delete maps call. Let > me redo the error handling so it's a bit more explicit. > > >> >>> out_maps: >>> xsk_delete_bpf_maps(xsk); >>> return err; >>> @@ -643,9 +596,7 @@ int xsk_socket__create(struct xsk_socket **xsk_ptr, const char *ifname, >>> goto out_mmap_tx; >>> } >>> >>> - xsk->qidconf_map_fd = -1; >>> - xsk->xsks_map_fd = -1; >>> - >>> + xsk->prog_fd = -1; >>> if (!(xsk->config.libbpf_flags & XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) { >>> err = xsk_setup_xdp_prog(xsk); >>> if (err) >>> @@ -708,8 +659,10 @@ void xsk_socket__delete(struct xsk_socket *xsk) >>> if (!xsk) >>> return; >>> >>> - xsk_clear_bpf_maps(xsk); >>> - xsk_delete_bpf_maps(xsk); >>> + if (xsk->prog_fd != -1) { >>> + xsk_delete_bpf_maps(xsk); >>> + close(xsk->prog_fd); >> >> Here, we use prog_fd != -1 to gate xsk_delete_bpf_maps(), which is >> confusing. I looked at the code for quite sometime, but still cannot >> confirm it is correct. > > See above reasoning - with INHIBIT_PROG_LOAD, there is no library-provided > program or maps, so cleanup actions are skipped. > -- > Jonathan Thanks for the explanation. Acked-by: Song Liu <songliubraving@fb.com>
On Sun, 2 Jun 2019 at 06:19, Jonathan Lemon <jonathan.lemon@gmail.com> wrote: > > > > On 1 Jun 2019, at 16:05, Song Liu wrote: > > >> On May 31, 2019, at 11:57 AM, Jonathan Lemon > >> <jonathan.lemon@gmail.com> wrote: > >> > >> Use the recent change to XSKMAP bpf_map_lookup_elem() to test if > >> there is a xsk present in the map instead of duplicating the work > >> with qidconf. > >> > >> Fix things so callers using XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD > >> bypass any internal bpf maps, so xsk_socket__{create|delete} works > >> properly. > >> > >> Signed-off-by: Jonathan Lemon <jonathan.lemon@gmail.com> > >> --- > >> tools/lib/bpf/xsk.c | 79 > >> +++++++++------------------------------------ > >> 1 file changed, 16 insertions(+), 63 deletions(-) > >> > >> diff --git a/tools/lib/bpf/xsk.c b/tools/lib/bpf/xsk.c > >> index 38667b62f1fe..7ce7494b5b50 100644 > >> --- a/tools/lib/bpf/xsk.c > >> +++ b/tools/lib/bpf/xsk.c > >> @@ -60,10 +60,8 @@ struct xsk_socket { > >> struct xsk_umem *umem; > >> struct xsk_socket_config config; > >> int fd; > >> - int xsks_map; > >> int ifindex; > >> int prog_fd; > >> - int qidconf_map_fd; > >> int xsks_map_fd; > >> __u32 queue_id; > >> char ifname[IFNAMSIZ]; > >> @@ -265,15 +263,11 @@ static int xsk_load_xdp_prog(struct xsk_socket > >> *xsk) > >> /* This is the C-program: > >> * SEC("xdp_sock") int xdp_sock_prog(struct xdp_md *ctx) > >> * { > >> - * int *qidconf, index = ctx->rx_queue_index; > >> + * int index = ctx->rx_queue_index; > >> * > >> * // A set entry here means that the correspnding queue_id > >> * // has an active AF_XDP socket bound to it. > >> - * qidconf = bpf_map_lookup_elem(&qidconf_map, &index); > >> - * if (!qidconf) > >> - * return XDP_ABORTED; > >> - * > >> - * if (*qidconf) > >> + * if (bpf_map_lookup_elem(&xsks_map, &index)) > >> * return bpf_redirect_map(&xsks_map, index, 0); > >> * > >> * return XDP_PASS; > >> @@ -286,15 +280,10 @@ static int xsk_load_xdp_prog(struct xsk_socket > >> *xsk) > >> BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_1, -4), > >> BPF_MOV64_REG(BPF_REG_2, BPF_REG_10), > >> BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), > >> - BPF_LD_MAP_FD(BPF_REG_1, xsk->qidconf_map_fd), > >> + BPF_LD_MAP_FD(BPF_REG_1, xsk->xsks_map_fd), > >> BPF_EMIT_CALL(BPF_FUNC_map_lookup_elem), > >> BPF_MOV64_REG(BPF_REG_1, BPF_REG_0), > >> - BPF_MOV32_IMM(BPF_REG_0, 0), > >> - /* if r1 == 0 goto +8 */ > >> - BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 8), > >> BPF_MOV32_IMM(BPF_REG_0, 2), > >> - /* r1 = *(u32 *)(r1 + 0) */ > >> - BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_1, 0), > >> /* if r1 == 0 goto +5 */ > >> BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 5), > >> /* r2 = *(u32 *)(r10 - 4) */ > >> @@ -366,18 +355,11 @@ static int xsk_create_bpf_maps(struct > >> xsk_socket *xsk) > >> if (max_queues < 0) > >> return max_queues; > >> > >> - fd = bpf_create_map_name(BPF_MAP_TYPE_ARRAY, "qidconf_map", > >> + fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map", > >> sizeof(int), sizeof(int), max_queues, 0); > >> if (fd < 0) > >> return fd; > >> - xsk->qidconf_map_fd = fd; > >> > >> - fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map", > >> - sizeof(int), sizeof(int), max_queues, 0); > >> - if (fd < 0) { > >> - close(xsk->qidconf_map_fd); > >> - return fd; > >> - } > >> xsk->xsks_map_fd = fd; > >> > >> return 0; > >> @@ -385,10 +367,8 @@ static int xsk_create_bpf_maps(struct xsk_socket > >> *xsk) > >> > >> static void xsk_delete_bpf_maps(struct xsk_socket *xsk) > >> { > >> - close(xsk->qidconf_map_fd); > >> + bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id); > >> close(xsk->xsks_map_fd); > >> - xsk->qidconf_map_fd = -1; > >> - xsk->xsks_map_fd = -1; > >> } > >> > >> static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) > >> @@ -417,10 +397,9 @@ static int xsk_lookup_bpf_maps(struct xsk_socket > >> *xsk) > >> if (err) > >> goto out_map_ids; > >> > >> - for (i = 0; i < prog_info.nr_map_ids; i++) { > >> - if (xsk->qidconf_map_fd != -1 && xsk->xsks_map_fd != -1) > >> - break; > >> + xsk->xsks_map_fd = -1; > >> > >> + for (i = 0; i < prog_info.nr_map_ids; i++) { > >> fd = bpf_map_get_fd_by_id(map_ids[i]); > >> if (fd < 0) > >> continue; > >> @@ -431,11 +410,6 @@ static int xsk_lookup_bpf_maps(struct xsk_socket > >> *xsk) > >> continue; > >> } > >> > >> - if (!strcmp(map_info.name, "qidconf_map")) { > >> - xsk->qidconf_map_fd = fd; > >> - continue; > >> - } > >> - > >> if (!strcmp(map_info.name, "xsks_map")) { > >> xsk->xsks_map_fd = fd; > >> continue; > >> @@ -445,40 +419,18 @@ static int xsk_lookup_bpf_maps(struct > >> xsk_socket *xsk) > >> } > >> > >> err = 0; > >> - if (xsk->qidconf_map_fd < 0 || xsk->xsks_map_fd < 0) { > >> + if (xsk->xsks_map_fd == -1) > >> err = -ENOENT; > >> - xsk_delete_bpf_maps(xsk); > >> - } > >> > >> out_map_ids: > >> free(map_ids); > >> return err; > >> } > >> > >> -static void xsk_clear_bpf_maps(struct xsk_socket *xsk) > >> -{ > >> - int qid = false; > >> - > >> - bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0); > >> - bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id); > >> -} > >> - > >> static int xsk_set_bpf_maps(struct xsk_socket *xsk) > >> { > >> - int qid = true, fd = xsk->fd, err; > >> - > >> - err = bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, > >> &qid, 0); > >> - if (err) > >> - goto out; > >> - > >> - err = bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, &fd, > >> 0); > >> - if (err) > >> - goto out; > >> - > >> - return 0; > >> -out: > >> - xsk_clear_bpf_maps(xsk); > >> - return err; > >> + return bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, > >> + &xsk->fd, 0); > >> } > >> > >> static int xsk_setup_xdp_prog(struct xsk_socket *xsk) > >> @@ -514,6 +466,7 @@ static int xsk_setup_xdp_prog(struct xsk_socket > >> *xsk) > >> > >> out_load: > >> close(xsk->prog_fd); > >> + xsk->prog_fd = -1; > > > > I found xsk->prog_fd confusing. Why do we need to set it here? > > I suppose this one isn't strictly required - I set it as a guard out of > habit. > xsk is (currently) immediately freed by the caller, so it can be > removed. > > > The main logic is: > > xsk->prog_fd = -1; > if (!(xsk->config.libbpf_flags & > XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) { > err = xsk_setup_xdp_prog(xsk); > > The user may pass INHIBIT_PROG_LOAD, which bypasses setting up the xdp > program > (and any maps associated with the program), allowing installation of a > custom > program. The cleanup behavior is then gated on prog_fd being -1, > > > > > I think we don't need to call xsk_delete_bpf_maps() in out_load path? > > Hmm, there's two out_load paths, but only one needs the delete maps > call. Let > me redo the error handling so it's a bit more explicit. > You're going for a V4? I took this series for a spin: Tested-by: Björn Töpel <bjorn.topel@intel.com> Very nice cleanup, Jonathan! Björn > > > > >> out_maps: > >> xsk_delete_bpf_maps(xsk); > >> return err; > >> @@ -643,9 +596,7 @@ int xsk_socket__create(struct xsk_socket > >> **xsk_ptr, const char *ifname, > >> goto out_mmap_tx; > >> } > >> > >> - xsk->qidconf_map_fd = -1; > >> - xsk->xsks_map_fd = -1; > >> - > >> + xsk->prog_fd = -1; > >> if (!(xsk->config.libbpf_flags & > >> XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) { > >> err = xsk_setup_xdp_prog(xsk); > >> if (err) > >> @@ -708,8 +659,10 @@ void xsk_socket__delete(struct xsk_socket *xsk) > >> if (!xsk) > >> return; > >> > >> - xsk_clear_bpf_maps(xsk); > >> - xsk_delete_bpf_maps(xsk); > >> + if (xsk->prog_fd != -1) { > >> + xsk_delete_bpf_maps(xsk); > >> + close(xsk->prog_fd); > > > > Here, we use prog_fd != -1 to gate xsk_delete_bpf_maps(), which is > > confusing. I looked at the code for quite sometime, but still cannot > > confirm it is correct. > > See above reasoning - with INHIBIT_PROG_LOAD, there is no > library-provided > program or maps, so cleanup actions are skipped. > -- > Jonathan
diff --git a/tools/lib/bpf/xsk.c b/tools/lib/bpf/xsk.c index 38667b62f1fe..7ce7494b5b50 100644 --- a/tools/lib/bpf/xsk.c +++ b/tools/lib/bpf/xsk.c @@ -60,10 +60,8 @@ struct xsk_socket { struct xsk_umem *umem; struct xsk_socket_config config; int fd; - int xsks_map; int ifindex; int prog_fd; - int qidconf_map_fd; int xsks_map_fd; __u32 queue_id; char ifname[IFNAMSIZ]; @@ -265,15 +263,11 @@ static int xsk_load_xdp_prog(struct xsk_socket *xsk) /* This is the C-program: * SEC("xdp_sock") int xdp_sock_prog(struct xdp_md *ctx) * { - * int *qidconf, index = ctx->rx_queue_index; + * int index = ctx->rx_queue_index; * * // A set entry here means that the correspnding queue_id * // has an active AF_XDP socket bound to it. - * qidconf = bpf_map_lookup_elem(&qidconf_map, &index); - * if (!qidconf) - * return XDP_ABORTED; - * - * if (*qidconf) + * if (bpf_map_lookup_elem(&xsks_map, &index)) * return bpf_redirect_map(&xsks_map, index, 0); * * return XDP_PASS; @@ -286,15 +280,10 @@ static int xsk_load_xdp_prog(struct xsk_socket *xsk) BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_1, -4), BPF_MOV64_REG(BPF_REG_2, BPF_REG_10), BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), - BPF_LD_MAP_FD(BPF_REG_1, xsk->qidconf_map_fd), + BPF_LD_MAP_FD(BPF_REG_1, xsk->xsks_map_fd), BPF_EMIT_CALL(BPF_FUNC_map_lookup_elem), BPF_MOV64_REG(BPF_REG_1, BPF_REG_0), - BPF_MOV32_IMM(BPF_REG_0, 0), - /* if r1 == 0 goto +8 */ - BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 8), BPF_MOV32_IMM(BPF_REG_0, 2), - /* r1 = *(u32 *)(r1 + 0) */ - BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_1, 0), /* if r1 == 0 goto +5 */ BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 5), /* r2 = *(u32 *)(r10 - 4) */ @@ -366,18 +355,11 @@ static int xsk_create_bpf_maps(struct xsk_socket *xsk) if (max_queues < 0) return max_queues; - fd = bpf_create_map_name(BPF_MAP_TYPE_ARRAY, "qidconf_map", + fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map", sizeof(int), sizeof(int), max_queues, 0); if (fd < 0) return fd; - xsk->qidconf_map_fd = fd; - fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map", - sizeof(int), sizeof(int), max_queues, 0); - if (fd < 0) { - close(xsk->qidconf_map_fd); - return fd; - } xsk->xsks_map_fd = fd; return 0; @@ -385,10 +367,8 @@ static int xsk_create_bpf_maps(struct xsk_socket *xsk) static void xsk_delete_bpf_maps(struct xsk_socket *xsk) { - close(xsk->qidconf_map_fd); + bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id); close(xsk->xsks_map_fd); - xsk->qidconf_map_fd = -1; - xsk->xsks_map_fd = -1; } static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) @@ -417,10 +397,9 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) if (err) goto out_map_ids; - for (i = 0; i < prog_info.nr_map_ids; i++) { - if (xsk->qidconf_map_fd != -1 && xsk->xsks_map_fd != -1) - break; + xsk->xsks_map_fd = -1; + for (i = 0; i < prog_info.nr_map_ids; i++) { fd = bpf_map_get_fd_by_id(map_ids[i]); if (fd < 0) continue; @@ -431,11 +410,6 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) continue; } - if (!strcmp(map_info.name, "qidconf_map")) { - xsk->qidconf_map_fd = fd; - continue; - } - if (!strcmp(map_info.name, "xsks_map")) { xsk->xsks_map_fd = fd; continue; @@ -445,40 +419,18 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk) } err = 0; - if (xsk->qidconf_map_fd < 0 || xsk->xsks_map_fd < 0) { + if (xsk->xsks_map_fd == -1) err = -ENOENT; - xsk_delete_bpf_maps(xsk); - } out_map_ids: free(map_ids); return err; } -static void xsk_clear_bpf_maps(struct xsk_socket *xsk) -{ - int qid = false; - - bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0); - bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id); -} - static int xsk_set_bpf_maps(struct xsk_socket *xsk) { - int qid = true, fd = xsk->fd, err; - - err = bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0); - if (err) - goto out; - - err = bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, &fd, 0); - if (err) - goto out; - - return 0; -out: - xsk_clear_bpf_maps(xsk); - return err; + return bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, + &xsk->fd, 0); } static int xsk_setup_xdp_prog(struct xsk_socket *xsk) @@ -514,6 +466,7 @@ static int xsk_setup_xdp_prog(struct xsk_socket *xsk) out_load: close(xsk->prog_fd); + xsk->prog_fd = -1; out_maps: xsk_delete_bpf_maps(xsk); return err; @@ -643,9 +596,7 @@ int xsk_socket__create(struct xsk_socket **xsk_ptr, const char *ifname, goto out_mmap_tx; } - xsk->qidconf_map_fd = -1; - xsk->xsks_map_fd = -1; - + xsk->prog_fd = -1; if (!(xsk->config.libbpf_flags & XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) { err = xsk_setup_xdp_prog(xsk); if (err) @@ -708,8 +659,10 @@ void xsk_socket__delete(struct xsk_socket *xsk) if (!xsk) return; - xsk_clear_bpf_maps(xsk); - xsk_delete_bpf_maps(xsk); + if (xsk->prog_fd != -1) { + xsk_delete_bpf_maps(xsk); + close(xsk->prog_fd); + } optlen = sizeof(off); err = getsockopt(xsk->fd, SOL_XDP, XDP_MMAP_OFFSETS, &off, &optlen);
Use the recent change to XSKMAP bpf_map_lookup_elem() to test if there is a xsk present in the map instead of duplicating the work with qidconf. Fix things so callers using XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD bypass any internal bpf maps, so xsk_socket__{create|delete} works properly. Signed-off-by: Jonathan Lemon <jonathan.lemon@gmail.com> --- tools/lib/bpf/xsk.c | 79 +++++++++------------------------------------ 1 file changed, 16 insertions(+), 63 deletions(-)