Skip to content

Commit

Permalink
bpf: minimize number of allocated lsm slots per program
Browse files Browse the repository at this point in the history
Previous patch adds 1:1 mapping between all 211 LSM hooks
and bpf_cgroup program array. Instead of reserving a slot per
possible hook, reserve 10 slots per cgroup for lsm programs.
Those slots are dynamically allocated on demand and reclaimed.

struct cgroup_bpf {
	struct bpf_prog_array *    effective[33];        /*     0   264 */
	/* --- cacheline 4 boundary (256 bytes) was 8 bytes ago --- */
	struct hlist_head          progs[33];            /*   264   264 */
	/* --- cacheline 8 boundary (512 bytes) was 16 bytes ago --- */
	u8                         flags[33];            /*   528    33 */

	/* XXX 7 bytes hole, try to pack */

	struct list_head           storages;             /*   568    16 */
	/* --- cacheline 9 boundary (576 bytes) was 8 bytes ago --- */
	struct bpf_prog_array *    inactive;             /*   584     8 */
	struct percpu_ref          refcnt;               /*   592    16 */
	struct work_struct         release_work;         /*   608    72 */

	/* size: 680, cachelines: 11, members: 7 */
	/* sum members: 673, holes: 1, sum holes: 7 */
	/* last cacheline: 40 bytes */
};

Move 'ifdef CONFIG_CGROUP_BPF' to expose CGROUP_BPF_ATTACH_TYPE_INVALID
to non-cgroup (core) parts.

Signed-off-by: Stanislav Fomichev <sdf@google.com>
  • Loading branch information
fomichev authored and intel-lab-lkp committed Jun 10, 2022
1 parent d4c2717 commit ca53950
Show file tree
Hide file tree
Showing 8 changed files with 66 additions and 24 deletions.
3 changes: 2 additions & 1 deletion include/linux/bpf-cgroup-defs.h
Expand Up @@ -11,7 +11,8 @@
struct bpf_prog_array;

#ifdef CONFIG_BPF_LSM
#define CGROUP_LSM_NUM 211 /* will be addressed in the next patch */
/* Maximum number of concurrently attachable per-cgroup LSM hooks. */
#define CGROUP_LSM_NUM 10
#else
#define CGROUP_LSM_NUM 0
#endif
Expand Down
4 changes: 3 additions & 1 deletion include/linux/bpf.h
Expand Up @@ -2407,7 +2407,6 @@ int bpf_arch_text_invalidate(void *dst, size_t len);

struct btf_id_set;
bool btf_id_set_contains(const struct btf_id_set *set, u32 id);
int btf_id_set_index(const struct btf_id_set *set, u32 id);

#define MAX_BPRINTF_VARARGS 12

Expand Down Expand Up @@ -2444,4 +2443,7 @@ void bpf_dynptr_init(struct bpf_dynptr_kern *ptr, void *data,
void bpf_dynptr_set_null(struct bpf_dynptr_kern *ptr);
int bpf_dynptr_check_size(u32 size);

void bpf_cgroup_atype_put(int cgroup_atype);
void bpf_cgroup_atype_get(u32 attach_btf_id, int cgroup_atype);

#endif /* _LINUX_BPF_H */
6 changes: 0 additions & 6 deletions include/linux/bpf_lsm.h
Expand Up @@ -42,7 +42,6 @@ extern const struct bpf_func_proto bpf_inode_storage_get_proto;
extern const struct bpf_func_proto bpf_inode_storage_delete_proto;
void bpf_inode_storage_free(struct inode *inode);

int bpf_lsm_hook_idx(u32 btf_id);
void bpf_lsm_find_cgroup_shim(const struct bpf_prog *prog, bpf_func_t *bpf_func);

#else /* !CONFIG_BPF_LSM */
Expand Down Expand Up @@ -73,11 +72,6 @@ static inline void bpf_lsm_find_cgroup_shim(const struct bpf_prog *prog,
{
}

static inline int bpf_lsm_hook_idx(u32 btf_id)
{
return -EINVAL;
}

#endif /* CONFIG_BPF_LSM */

#endif /* _LINUX_BPF_LSM_H */
5 changes: 0 additions & 5 deletions kernel/bpf/bpf_lsm.c
Expand Up @@ -69,11 +69,6 @@ void bpf_lsm_find_cgroup_shim(const struct bpf_prog *prog,
*bpf_func = __cgroup_bpf_run_lsm_current;
}

int bpf_lsm_hook_idx(u32 btf_id)
{
return btf_id_set_index(&bpf_lsm_hooks, btf_id);
}

int bpf_lsm_verify_prog(struct bpf_verifier_log *vlog,
const struct bpf_prog *prog)
{
Expand Down
10 changes: 0 additions & 10 deletions kernel/bpf/btf.c
Expand Up @@ -6842,16 +6842,6 @@ static int btf_id_cmp_func(const void *a, const void *b)
return *pa - *pb;
}

int btf_id_set_index(const struct btf_id_set *set, u32 id)
{
const u32 *p;

p = bsearch(&id, set->ids, set->cnt, sizeof(u32), btf_id_cmp_func);
if (!p)
return -1;
return p - set->ids;
}

bool btf_id_set_contains(const struct btf_id_set *set, u32 id)
{
return bsearch(&id, set->ids, set->cnt, sizeof(u32), btf_id_cmp_func) != NULL;
Expand Down
54 changes: 53 additions & 1 deletion kernel/bpf/cgroup.c
Expand Up @@ -128,12 +128,56 @@ unsigned int __cgroup_bpf_run_lsm_current(const void *ctx,
}

#ifdef CONFIG_BPF_LSM
struct cgroup_lsm_atype {
u32 attach_btf_id;
int refcnt;
};

static struct cgroup_lsm_atype cgroup_lsm_atype[CGROUP_LSM_NUM];

static enum cgroup_bpf_attach_type
bpf_cgroup_atype_find(enum bpf_attach_type attach_type, u32 attach_btf_id)
{
int i;

lockdep_assert_held(&cgroup_mutex);

if (attach_type != BPF_LSM_CGROUP)
return to_cgroup_bpf_attach_type(attach_type);
return CGROUP_LSM_START + bpf_lsm_hook_idx(attach_btf_id);

for (i = 0; i < ARRAY_SIZE(cgroup_lsm_atype); i++)
if (cgroup_lsm_atype[i].attach_btf_id == attach_btf_id)
return CGROUP_LSM_START + i;

for (i = 0; i < ARRAY_SIZE(cgroup_lsm_atype); i++)
if (cgroup_lsm_atype[i].attach_btf_id == 0)
return CGROUP_LSM_START + i;

return -E2BIG;

}

void bpf_cgroup_atype_get(u32 attach_btf_id, int cgroup_atype)
{
int i = cgroup_atype - CGROUP_LSM_START;

lockdep_assert_held(&cgroup_mutex);

WARN_ON_ONCE(cgroup_lsm_atype[i].attach_btf_id &&
cgroup_lsm_atype[i].attach_btf_id != attach_btf_id);

cgroup_lsm_atype[i].attach_btf_id = attach_btf_id;
cgroup_lsm_atype[i].refcnt++;
}

void bpf_cgroup_atype_put(int cgroup_atype)
{
int i = cgroup_atype - CGROUP_LSM_START;

mutex_lock(&cgroup_mutex);
if (--cgroup_lsm_atype[i].refcnt <= 0)
cgroup_lsm_atype[i].attach_btf_id = 0;
mutex_unlock(&cgroup_mutex);
}
#else
static enum cgroup_bpf_attach_type
Expand All @@ -143,6 +187,14 @@ bpf_cgroup_atype_find(enum bpf_attach_type attach_type, u32 attach_btf_id)
return to_cgroup_bpf_attach_type(attach_type);
return -EOPNOTSUPP;
}

void bpf_cgroup_atype_get(u32 attach_btf_id, int cgroup_atype)
{
}

void bpf_cgroup_atype_put(int cgroup_atype)
{
}
#endif /* CONFIG_BPF_LSM */

void cgroup_bpf_offline(struct cgroup *cgrp)
Expand Down
7 changes: 7 additions & 0 deletions kernel/bpf/core.c
Expand Up @@ -107,6 +107,9 @@ struct bpf_prog *bpf_prog_alloc_no_stats(unsigned int size, gfp_t gfp_extra_flag
fp->aux->prog = fp;
fp->jit_requested = ebpf_jit_enabled();
fp->blinding_requested = bpf_jit_blinding_enabled(fp);
#ifdef CONFIG_CGROUP_BPF
aux->cgroup_atype = CGROUP_BPF_ATTACH_TYPE_INVALID;
#endif

INIT_LIST_HEAD_RCU(&fp->aux->ksym.lnode);
mutex_init(&fp->aux->used_maps_mutex);
Expand Down Expand Up @@ -2554,6 +2557,10 @@ static void bpf_prog_free_deferred(struct work_struct *work)
aux = container_of(work, struct bpf_prog_aux, work);
#ifdef CONFIG_BPF_SYSCALL
bpf_free_kfunc_btf_tab(aux->kfunc_btf_tab);
#endif
#ifdef CONFIG_CGROUP_BPF
if (aux->cgroup_atype != CGROUP_BPF_ATTACH_TYPE_INVALID)
bpf_cgroup_atype_put(aux->cgroup_atype);
#endif
bpf_free_used_maps(aux);
bpf_free_used_btfs(aux);
Expand Down
1 change: 1 addition & 0 deletions kernel/bpf/trampoline.c
Expand Up @@ -555,6 +555,7 @@ static struct bpf_shim_tramp_link *cgroup_shim_alloc(const struct bpf_prog *prog
bpf_prog_inc(p);
bpf_link_init(&shim_link->link.link, BPF_LINK_TYPE_UNSPEC,
&bpf_shim_tramp_link_lops, p);
bpf_cgroup_atype_get(p->aux->attach_btf_id, cgroup_atype);

return shim_link;
}
Expand Down

0 comments on commit ca53950

Please sign in to comment.