xref: /OK3568_Linux_fs/kernel/kernel/bpf/inode.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * Minimal file system backend for holding eBPF maps and programs,
4*4882a593Smuzhiyun  * used by bpf(2) object pinning.
5*4882a593Smuzhiyun  *
6*4882a593Smuzhiyun  * Authors:
7*4882a593Smuzhiyun  *
8*4882a593Smuzhiyun  *	Daniel Borkmann <daniel@iogearbox.net>
9*4882a593Smuzhiyun  */
10*4882a593Smuzhiyun 
11*4882a593Smuzhiyun #include <linux/init.h>
12*4882a593Smuzhiyun #include <linux/magic.h>
13*4882a593Smuzhiyun #include <linux/major.h>
14*4882a593Smuzhiyun #include <linux/mount.h>
15*4882a593Smuzhiyun #include <linux/namei.h>
16*4882a593Smuzhiyun #include <linux/fs.h>
17*4882a593Smuzhiyun #include <linux/fs_context.h>
18*4882a593Smuzhiyun #include <linux/fs_parser.h>
19*4882a593Smuzhiyun #include <linux/kdev_t.h>
20*4882a593Smuzhiyun #include <linux/filter.h>
21*4882a593Smuzhiyun #include <linux/bpf.h>
22*4882a593Smuzhiyun #include <linux/bpf_trace.h>
23*4882a593Smuzhiyun #include "preload/bpf_preload.h"
24*4882a593Smuzhiyun 
25*4882a593Smuzhiyun enum bpf_type {
26*4882a593Smuzhiyun 	BPF_TYPE_UNSPEC	= 0,
27*4882a593Smuzhiyun 	BPF_TYPE_PROG,
28*4882a593Smuzhiyun 	BPF_TYPE_MAP,
29*4882a593Smuzhiyun 	BPF_TYPE_LINK,
30*4882a593Smuzhiyun };
31*4882a593Smuzhiyun 
bpf_any_get(void * raw,enum bpf_type type)32*4882a593Smuzhiyun static void *bpf_any_get(void *raw, enum bpf_type type)
33*4882a593Smuzhiyun {
34*4882a593Smuzhiyun 	switch (type) {
35*4882a593Smuzhiyun 	case BPF_TYPE_PROG:
36*4882a593Smuzhiyun 		bpf_prog_inc(raw);
37*4882a593Smuzhiyun 		break;
38*4882a593Smuzhiyun 	case BPF_TYPE_MAP:
39*4882a593Smuzhiyun 		bpf_map_inc_with_uref(raw);
40*4882a593Smuzhiyun 		break;
41*4882a593Smuzhiyun 	case BPF_TYPE_LINK:
42*4882a593Smuzhiyun 		bpf_link_inc(raw);
43*4882a593Smuzhiyun 		break;
44*4882a593Smuzhiyun 	default:
45*4882a593Smuzhiyun 		WARN_ON_ONCE(1);
46*4882a593Smuzhiyun 		break;
47*4882a593Smuzhiyun 	}
48*4882a593Smuzhiyun 
49*4882a593Smuzhiyun 	return raw;
50*4882a593Smuzhiyun }
51*4882a593Smuzhiyun 
bpf_any_put(void * raw,enum bpf_type type)52*4882a593Smuzhiyun static void bpf_any_put(void *raw, enum bpf_type type)
53*4882a593Smuzhiyun {
54*4882a593Smuzhiyun 	switch (type) {
55*4882a593Smuzhiyun 	case BPF_TYPE_PROG:
56*4882a593Smuzhiyun 		bpf_prog_put(raw);
57*4882a593Smuzhiyun 		break;
58*4882a593Smuzhiyun 	case BPF_TYPE_MAP:
59*4882a593Smuzhiyun 		bpf_map_put_with_uref(raw);
60*4882a593Smuzhiyun 		break;
61*4882a593Smuzhiyun 	case BPF_TYPE_LINK:
62*4882a593Smuzhiyun 		bpf_link_put(raw);
63*4882a593Smuzhiyun 		break;
64*4882a593Smuzhiyun 	default:
65*4882a593Smuzhiyun 		WARN_ON_ONCE(1);
66*4882a593Smuzhiyun 		break;
67*4882a593Smuzhiyun 	}
68*4882a593Smuzhiyun }
69*4882a593Smuzhiyun 
bpf_fd_probe_obj(u32 ufd,enum bpf_type * type)70*4882a593Smuzhiyun static void *bpf_fd_probe_obj(u32 ufd, enum bpf_type *type)
71*4882a593Smuzhiyun {
72*4882a593Smuzhiyun 	void *raw;
73*4882a593Smuzhiyun 
74*4882a593Smuzhiyun 	raw = bpf_map_get_with_uref(ufd);
75*4882a593Smuzhiyun 	if (!IS_ERR(raw)) {
76*4882a593Smuzhiyun 		*type = BPF_TYPE_MAP;
77*4882a593Smuzhiyun 		return raw;
78*4882a593Smuzhiyun 	}
79*4882a593Smuzhiyun 
80*4882a593Smuzhiyun 	raw = bpf_prog_get(ufd);
81*4882a593Smuzhiyun 	if (!IS_ERR(raw)) {
82*4882a593Smuzhiyun 		*type = BPF_TYPE_PROG;
83*4882a593Smuzhiyun 		return raw;
84*4882a593Smuzhiyun 	}
85*4882a593Smuzhiyun 
86*4882a593Smuzhiyun 	raw = bpf_link_get_from_fd(ufd);
87*4882a593Smuzhiyun 	if (!IS_ERR(raw)) {
88*4882a593Smuzhiyun 		*type = BPF_TYPE_LINK;
89*4882a593Smuzhiyun 		return raw;
90*4882a593Smuzhiyun 	}
91*4882a593Smuzhiyun 
92*4882a593Smuzhiyun 	return ERR_PTR(-EINVAL);
93*4882a593Smuzhiyun }
94*4882a593Smuzhiyun 
95*4882a593Smuzhiyun static const struct inode_operations bpf_dir_iops;
96*4882a593Smuzhiyun 
97*4882a593Smuzhiyun static const struct inode_operations bpf_prog_iops = { };
98*4882a593Smuzhiyun static const struct inode_operations bpf_map_iops  = { };
99*4882a593Smuzhiyun static const struct inode_operations bpf_link_iops  = { };
100*4882a593Smuzhiyun 
bpf_get_inode(struct super_block * sb,const struct inode * dir,umode_t mode)101*4882a593Smuzhiyun static struct inode *bpf_get_inode(struct super_block *sb,
102*4882a593Smuzhiyun 				   const struct inode *dir,
103*4882a593Smuzhiyun 				   umode_t mode)
104*4882a593Smuzhiyun {
105*4882a593Smuzhiyun 	struct inode *inode;
106*4882a593Smuzhiyun 
107*4882a593Smuzhiyun 	switch (mode & S_IFMT) {
108*4882a593Smuzhiyun 	case S_IFDIR:
109*4882a593Smuzhiyun 	case S_IFREG:
110*4882a593Smuzhiyun 	case S_IFLNK:
111*4882a593Smuzhiyun 		break;
112*4882a593Smuzhiyun 	default:
113*4882a593Smuzhiyun 		return ERR_PTR(-EINVAL);
114*4882a593Smuzhiyun 	}
115*4882a593Smuzhiyun 
116*4882a593Smuzhiyun 	inode = new_inode(sb);
117*4882a593Smuzhiyun 	if (!inode)
118*4882a593Smuzhiyun 		return ERR_PTR(-ENOSPC);
119*4882a593Smuzhiyun 
120*4882a593Smuzhiyun 	inode->i_ino = get_next_ino();
121*4882a593Smuzhiyun 	inode->i_atime = current_time(inode);
122*4882a593Smuzhiyun 	inode->i_mtime = inode->i_atime;
123*4882a593Smuzhiyun 	inode->i_ctime = inode->i_atime;
124*4882a593Smuzhiyun 
125*4882a593Smuzhiyun 	inode_init_owner(inode, dir, mode);
126*4882a593Smuzhiyun 
127*4882a593Smuzhiyun 	return inode;
128*4882a593Smuzhiyun }
129*4882a593Smuzhiyun 
bpf_inode_type(const struct inode * inode,enum bpf_type * type)130*4882a593Smuzhiyun static int bpf_inode_type(const struct inode *inode, enum bpf_type *type)
131*4882a593Smuzhiyun {
132*4882a593Smuzhiyun 	*type = BPF_TYPE_UNSPEC;
133*4882a593Smuzhiyun 	if (inode->i_op == &bpf_prog_iops)
134*4882a593Smuzhiyun 		*type = BPF_TYPE_PROG;
135*4882a593Smuzhiyun 	else if (inode->i_op == &bpf_map_iops)
136*4882a593Smuzhiyun 		*type = BPF_TYPE_MAP;
137*4882a593Smuzhiyun 	else if (inode->i_op == &bpf_link_iops)
138*4882a593Smuzhiyun 		*type = BPF_TYPE_LINK;
139*4882a593Smuzhiyun 	else
140*4882a593Smuzhiyun 		return -EACCES;
141*4882a593Smuzhiyun 
142*4882a593Smuzhiyun 	return 0;
143*4882a593Smuzhiyun }
144*4882a593Smuzhiyun 
bpf_dentry_finalize(struct dentry * dentry,struct inode * inode,struct inode * dir)145*4882a593Smuzhiyun static void bpf_dentry_finalize(struct dentry *dentry, struct inode *inode,
146*4882a593Smuzhiyun 				struct inode *dir)
147*4882a593Smuzhiyun {
148*4882a593Smuzhiyun 	d_instantiate(dentry, inode);
149*4882a593Smuzhiyun 	dget(dentry);
150*4882a593Smuzhiyun 
151*4882a593Smuzhiyun 	dir->i_mtime = current_time(dir);
152*4882a593Smuzhiyun 	dir->i_ctime = dir->i_mtime;
153*4882a593Smuzhiyun }
154*4882a593Smuzhiyun 
bpf_mkdir(struct inode * dir,struct dentry * dentry,umode_t mode)155*4882a593Smuzhiyun static int bpf_mkdir(struct inode *dir, struct dentry *dentry, umode_t mode)
156*4882a593Smuzhiyun {
157*4882a593Smuzhiyun 	struct inode *inode;
158*4882a593Smuzhiyun 
159*4882a593Smuzhiyun 	inode = bpf_get_inode(dir->i_sb, dir, mode | S_IFDIR);
160*4882a593Smuzhiyun 	if (IS_ERR(inode))
161*4882a593Smuzhiyun 		return PTR_ERR(inode);
162*4882a593Smuzhiyun 
163*4882a593Smuzhiyun 	inode->i_op = &bpf_dir_iops;
164*4882a593Smuzhiyun 	inode->i_fop = &simple_dir_operations;
165*4882a593Smuzhiyun 
166*4882a593Smuzhiyun 	inc_nlink(inode);
167*4882a593Smuzhiyun 	inc_nlink(dir);
168*4882a593Smuzhiyun 
169*4882a593Smuzhiyun 	bpf_dentry_finalize(dentry, inode, dir);
170*4882a593Smuzhiyun 	return 0;
171*4882a593Smuzhiyun }
172*4882a593Smuzhiyun 
173*4882a593Smuzhiyun struct map_iter {
174*4882a593Smuzhiyun 	void *key;
175*4882a593Smuzhiyun 	bool done;
176*4882a593Smuzhiyun };
177*4882a593Smuzhiyun 
map_iter(struct seq_file * m)178*4882a593Smuzhiyun static struct map_iter *map_iter(struct seq_file *m)
179*4882a593Smuzhiyun {
180*4882a593Smuzhiyun 	return m->private;
181*4882a593Smuzhiyun }
182*4882a593Smuzhiyun 
seq_file_to_map(struct seq_file * m)183*4882a593Smuzhiyun static struct bpf_map *seq_file_to_map(struct seq_file *m)
184*4882a593Smuzhiyun {
185*4882a593Smuzhiyun 	return file_inode(m->file)->i_private;
186*4882a593Smuzhiyun }
187*4882a593Smuzhiyun 
map_iter_free(struct map_iter * iter)188*4882a593Smuzhiyun static void map_iter_free(struct map_iter *iter)
189*4882a593Smuzhiyun {
190*4882a593Smuzhiyun 	if (iter) {
191*4882a593Smuzhiyun 		kfree(iter->key);
192*4882a593Smuzhiyun 		kfree(iter);
193*4882a593Smuzhiyun 	}
194*4882a593Smuzhiyun }
195*4882a593Smuzhiyun 
map_iter_alloc(struct bpf_map * map)196*4882a593Smuzhiyun static struct map_iter *map_iter_alloc(struct bpf_map *map)
197*4882a593Smuzhiyun {
198*4882a593Smuzhiyun 	struct map_iter *iter;
199*4882a593Smuzhiyun 
200*4882a593Smuzhiyun 	iter = kzalloc(sizeof(*iter), GFP_KERNEL | __GFP_NOWARN);
201*4882a593Smuzhiyun 	if (!iter)
202*4882a593Smuzhiyun 		goto error;
203*4882a593Smuzhiyun 
204*4882a593Smuzhiyun 	iter->key = kzalloc(map->key_size, GFP_KERNEL | __GFP_NOWARN);
205*4882a593Smuzhiyun 	if (!iter->key)
206*4882a593Smuzhiyun 		goto error;
207*4882a593Smuzhiyun 
208*4882a593Smuzhiyun 	return iter;
209*4882a593Smuzhiyun 
210*4882a593Smuzhiyun error:
211*4882a593Smuzhiyun 	map_iter_free(iter);
212*4882a593Smuzhiyun 	return NULL;
213*4882a593Smuzhiyun }
214*4882a593Smuzhiyun 
map_seq_next(struct seq_file * m,void * v,loff_t * pos)215*4882a593Smuzhiyun static void *map_seq_next(struct seq_file *m, void *v, loff_t *pos)
216*4882a593Smuzhiyun {
217*4882a593Smuzhiyun 	struct bpf_map *map = seq_file_to_map(m);
218*4882a593Smuzhiyun 	void *key = map_iter(m)->key;
219*4882a593Smuzhiyun 	void *prev_key;
220*4882a593Smuzhiyun 
221*4882a593Smuzhiyun 	(*pos)++;
222*4882a593Smuzhiyun 	if (map_iter(m)->done)
223*4882a593Smuzhiyun 		return NULL;
224*4882a593Smuzhiyun 
225*4882a593Smuzhiyun 	if (unlikely(v == SEQ_START_TOKEN))
226*4882a593Smuzhiyun 		prev_key = NULL;
227*4882a593Smuzhiyun 	else
228*4882a593Smuzhiyun 		prev_key = key;
229*4882a593Smuzhiyun 
230*4882a593Smuzhiyun 	rcu_read_lock();
231*4882a593Smuzhiyun 	if (map->ops->map_get_next_key(map, prev_key, key)) {
232*4882a593Smuzhiyun 		map_iter(m)->done = true;
233*4882a593Smuzhiyun 		key = NULL;
234*4882a593Smuzhiyun 	}
235*4882a593Smuzhiyun 	rcu_read_unlock();
236*4882a593Smuzhiyun 	return key;
237*4882a593Smuzhiyun }
238*4882a593Smuzhiyun 
map_seq_start(struct seq_file * m,loff_t * pos)239*4882a593Smuzhiyun static void *map_seq_start(struct seq_file *m, loff_t *pos)
240*4882a593Smuzhiyun {
241*4882a593Smuzhiyun 	if (map_iter(m)->done)
242*4882a593Smuzhiyun 		return NULL;
243*4882a593Smuzhiyun 
244*4882a593Smuzhiyun 	return *pos ? map_iter(m)->key : SEQ_START_TOKEN;
245*4882a593Smuzhiyun }
246*4882a593Smuzhiyun 
map_seq_stop(struct seq_file * m,void * v)247*4882a593Smuzhiyun static void map_seq_stop(struct seq_file *m, void *v)
248*4882a593Smuzhiyun {
249*4882a593Smuzhiyun }
250*4882a593Smuzhiyun 
map_seq_show(struct seq_file * m,void * v)251*4882a593Smuzhiyun static int map_seq_show(struct seq_file *m, void *v)
252*4882a593Smuzhiyun {
253*4882a593Smuzhiyun 	struct bpf_map *map = seq_file_to_map(m);
254*4882a593Smuzhiyun 	void *key = map_iter(m)->key;
255*4882a593Smuzhiyun 
256*4882a593Smuzhiyun 	if (unlikely(v == SEQ_START_TOKEN)) {
257*4882a593Smuzhiyun 		seq_puts(m, "# WARNING!! The output is for debug purpose only\n");
258*4882a593Smuzhiyun 		seq_puts(m, "# WARNING!! The output format will change\n");
259*4882a593Smuzhiyun 	} else {
260*4882a593Smuzhiyun 		map->ops->map_seq_show_elem(map, key, m);
261*4882a593Smuzhiyun 	}
262*4882a593Smuzhiyun 
263*4882a593Smuzhiyun 	return 0;
264*4882a593Smuzhiyun }
265*4882a593Smuzhiyun 
266*4882a593Smuzhiyun static const struct seq_operations bpffs_map_seq_ops = {
267*4882a593Smuzhiyun 	.start	= map_seq_start,
268*4882a593Smuzhiyun 	.next	= map_seq_next,
269*4882a593Smuzhiyun 	.show	= map_seq_show,
270*4882a593Smuzhiyun 	.stop	= map_seq_stop,
271*4882a593Smuzhiyun };
272*4882a593Smuzhiyun 
bpffs_map_open(struct inode * inode,struct file * file)273*4882a593Smuzhiyun static int bpffs_map_open(struct inode *inode, struct file *file)
274*4882a593Smuzhiyun {
275*4882a593Smuzhiyun 	struct bpf_map *map = inode->i_private;
276*4882a593Smuzhiyun 	struct map_iter *iter;
277*4882a593Smuzhiyun 	struct seq_file *m;
278*4882a593Smuzhiyun 	int err;
279*4882a593Smuzhiyun 
280*4882a593Smuzhiyun 	iter = map_iter_alloc(map);
281*4882a593Smuzhiyun 	if (!iter)
282*4882a593Smuzhiyun 		return -ENOMEM;
283*4882a593Smuzhiyun 
284*4882a593Smuzhiyun 	err = seq_open(file, &bpffs_map_seq_ops);
285*4882a593Smuzhiyun 	if (err) {
286*4882a593Smuzhiyun 		map_iter_free(iter);
287*4882a593Smuzhiyun 		return err;
288*4882a593Smuzhiyun 	}
289*4882a593Smuzhiyun 
290*4882a593Smuzhiyun 	m = file->private_data;
291*4882a593Smuzhiyun 	m->private = iter;
292*4882a593Smuzhiyun 
293*4882a593Smuzhiyun 	return 0;
294*4882a593Smuzhiyun }
295*4882a593Smuzhiyun 
bpffs_map_release(struct inode * inode,struct file * file)296*4882a593Smuzhiyun static int bpffs_map_release(struct inode *inode, struct file *file)
297*4882a593Smuzhiyun {
298*4882a593Smuzhiyun 	struct seq_file *m = file->private_data;
299*4882a593Smuzhiyun 
300*4882a593Smuzhiyun 	map_iter_free(map_iter(m));
301*4882a593Smuzhiyun 
302*4882a593Smuzhiyun 	return seq_release(inode, file);
303*4882a593Smuzhiyun }
304*4882a593Smuzhiyun 
305*4882a593Smuzhiyun /* bpffs_map_fops should only implement the basic
306*4882a593Smuzhiyun  * read operation for a BPF map.  The purpose is to
307*4882a593Smuzhiyun  * provide a simple user intuitive way to do
308*4882a593Smuzhiyun  * "cat bpffs/pathto/a-pinned-map".
309*4882a593Smuzhiyun  *
310*4882a593Smuzhiyun  * Other operations (e.g. write, lookup...) should be realized by
311*4882a593Smuzhiyun  * the userspace tools (e.g. bpftool) through the
312*4882a593Smuzhiyun  * BPF_OBJ_GET_INFO_BY_FD and the map's lookup/update
313*4882a593Smuzhiyun  * interface.
314*4882a593Smuzhiyun  */
315*4882a593Smuzhiyun static const struct file_operations bpffs_map_fops = {
316*4882a593Smuzhiyun 	.open		= bpffs_map_open,
317*4882a593Smuzhiyun 	.read		= seq_read,
318*4882a593Smuzhiyun 	.release	= bpffs_map_release,
319*4882a593Smuzhiyun };
320*4882a593Smuzhiyun 
bpffs_obj_open(struct inode * inode,struct file * file)321*4882a593Smuzhiyun static int bpffs_obj_open(struct inode *inode, struct file *file)
322*4882a593Smuzhiyun {
323*4882a593Smuzhiyun 	return -EIO;
324*4882a593Smuzhiyun }
325*4882a593Smuzhiyun 
326*4882a593Smuzhiyun static const struct file_operations bpffs_obj_fops = {
327*4882a593Smuzhiyun 	.open		= bpffs_obj_open,
328*4882a593Smuzhiyun };
329*4882a593Smuzhiyun 
bpf_mkobj_ops(struct dentry * dentry,umode_t mode,void * raw,const struct inode_operations * iops,const struct file_operations * fops)330*4882a593Smuzhiyun static int bpf_mkobj_ops(struct dentry *dentry, umode_t mode, void *raw,
331*4882a593Smuzhiyun 			 const struct inode_operations *iops,
332*4882a593Smuzhiyun 			 const struct file_operations *fops)
333*4882a593Smuzhiyun {
334*4882a593Smuzhiyun 	struct inode *dir = dentry->d_parent->d_inode;
335*4882a593Smuzhiyun 	struct inode *inode = bpf_get_inode(dir->i_sb, dir, mode);
336*4882a593Smuzhiyun 	if (IS_ERR(inode))
337*4882a593Smuzhiyun 		return PTR_ERR(inode);
338*4882a593Smuzhiyun 
339*4882a593Smuzhiyun 	inode->i_op = iops;
340*4882a593Smuzhiyun 	inode->i_fop = fops;
341*4882a593Smuzhiyun 	inode->i_private = raw;
342*4882a593Smuzhiyun 
343*4882a593Smuzhiyun 	bpf_dentry_finalize(dentry, inode, dir);
344*4882a593Smuzhiyun 	return 0;
345*4882a593Smuzhiyun }
346*4882a593Smuzhiyun 
bpf_mkprog(struct dentry * dentry,umode_t mode,void * arg)347*4882a593Smuzhiyun static int bpf_mkprog(struct dentry *dentry, umode_t mode, void *arg)
348*4882a593Smuzhiyun {
349*4882a593Smuzhiyun 	return bpf_mkobj_ops(dentry, mode, arg, &bpf_prog_iops,
350*4882a593Smuzhiyun 			     &bpffs_obj_fops);
351*4882a593Smuzhiyun }
352*4882a593Smuzhiyun 
bpf_mkmap(struct dentry * dentry,umode_t mode,void * arg)353*4882a593Smuzhiyun static int bpf_mkmap(struct dentry *dentry, umode_t mode, void *arg)
354*4882a593Smuzhiyun {
355*4882a593Smuzhiyun 	struct bpf_map *map = arg;
356*4882a593Smuzhiyun 
357*4882a593Smuzhiyun 	return bpf_mkobj_ops(dentry, mode, arg, &bpf_map_iops,
358*4882a593Smuzhiyun 			     bpf_map_support_seq_show(map) ?
359*4882a593Smuzhiyun 			     &bpffs_map_fops : &bpffs_obj_fops);
360*4882a593Smuzhiyun }
361*4882a593Smuzhiyun 
bpf_mklink(struct dentry * dentry,umode_t mode,void * arg)362*4882a593Smuzhiyun static int bpf_mklink(struct dentry *dentry, umode_t mode, void *arg)
363*4882a593Smuzhiyun {
364*4882a593Smuzhiyun 	struct bpf_link *link = arg;
365*4882a593Smuzhiyun 
366*4882a593Smuzhiyun 	return bpf_mkobj_ops(dentry, mode, arg, &bpf_link_iops,
367*4882a593Smuzhiyun 			     bpf_link_is_iter(link) ?
368*4882a593Smuzhiyun 			     &bpf_iter_fops : &bpffs_obj_fops);
369*4882a593Smuzhiyun }
370*4882a593Smuzhiyun 
371*4882a593Smuzhiyun static struct dentry *
bpf_lookup(struct inode * dir,struct dentry * dentry,unsigned flags)372*4882a593Smuzhiyun bpf_lookup(struct inode *dir, struct dentry *dentry, unsigned flags)
373*4882a593Smuzhiyun {
374*4882a593Smuzhiyun 	/* Dots in names (e.g. "/sys/fs/bpf/foo.bar") are reserved for future
375*4882a593Smuzhiyun 	 * extensions. That allows popoulate_bpffs() create special files.
376*4882a593Smuzhiyun 	 */
377*4882a593Smuzhiyun 	if ((dir->i_mode & S_IALLUGO) &&
378*4882a593Smuzhiyun 	    strchr(dentry->d_name.name, '.'))
379*4882a593Smuzhiyun 		return ERR_PTR(-EPERM);
380*4882a593Smuzhiyun 
381*4882a593Smuzhiyun 	return simple_lookup(dir, dentry, flags);
382*4882a593Smuzhiyun }
383*4882a593Smuzhiyun 
bpf_symlink(struct inode * dir,struct dentry * dentry,const char * target)384*4882a593Smuzhiyun static int bpf_symlink(struct inode *dir, struct dentry *dentry,
385*4882a593Smuzhiyun 		       const char *target)
386*4882a593Smuzhiyun {
387*4882a593Smuzhiyun 	char *link = kstrdup(target, GFP_USER | __GFP_NOWARN);
388*4882a593Smuzhiyun 	struct inode *inode;
389*4882a593Smuzhiyun 
390*4882a593Smuzhiyun 	if (!link)
391*4882a593Smuzhiyun 		return -ENOMEM;
392*4882a593Smuzhiyun 
393*4882a593Smuzhiyun 	inode = bpf_get_inode(dir->i_sb, dir, S_IRWXUGO | S_IFLNK);
394*4882a593Smuzhiyun 	if (IS_ERR(inode)) {
395*4882a593Smuzhiyun 		kfree(link);
396*4882a593Smuzhiyun 		return PTR_ERR(inode);
397*4882a593Smuzhiyun 	}
398*4882a593Smuzhiyun 
399*4882a593Smuzhiyun 	inode->i_op = &simple_symlink_inode_operations;
400*4882a593Smuzhiyun 	inode->i_link = link;
401*4882a593Smuzhiyun 
402*4882a593Smuzhiyun 	bpf_dentry_finalize(dentry, inode, dir);
403*4882a593Smuzhiyun 	return 0;
404*4882a593Smuzhiyun }
405*4882a593Smuzhiyun 
406*4882a593Smuzhiyun static const struct inode_operations bpf_dir_iops = {
407*4882a593Smuzhiyun 	.lookup		= bpf_lookup,
408*4882a593Smuzhiyun 	.mkdir		= bpf_mkdir,
409*4882a593Smuzhiyun 	.symlink	= bpf_symlink,
410*4882a593Smuzhiyun 	.rmdir		= simple_rmdir,
411*4882a593Smuzhiyun 	.rename		= simple_rename,
412*4882a593Smuzhiyun 	.link		= simple_link,
413*4882a593Smuzhiyun 	.unlink		= simple_unlink,
414*4882a593Smuzhiyun };
415*4882a593Smuzhiyun 
416*4882a593Smuzhiyun /* pin iterator link into bpffs */
bpf_iter_link_pin_kernel(struct dentry * parent,const char * name,struct bpf_link * link)417*4882a593Smuzhiyun static int bpf_iter_link_pin_kernel(struct dentry *parent,
418*4882a593Smuzhiyun 				    const char *name, struct bpf_link *link)
419*4882a593Smuzhiyun {
420*4882a593Smuzhiyun 	umode_t mode = S_IFREG | S_IRUSR;
421*4882a593Smuzhiyun 	struct dentry *dentry;
422*4882a593Smuzhiyun 	int ret;
423*4882a593Smuzhiyun 
424*4882a593Smuzhiyun 	inode_lock(parent->d_inode);
425*4882a593Smuzhiyun 	dentry = lookup_one_len(name, parent, strlen(name));
426*4882a593Smuzhiyun 	if (IS_ERR(dentry)) {
427*4882a593Smuzhiyun 		inode_unlock(parent->d_inode);
428*4882a593Smuzhiyun 		return PTR_ERR(dentry);
429*4882a593Smuzhiyun 	}
430*4882a593Smuzhiyun 	ret = bpf_mkobj_ops(dentry, mode, link, &bpf_link_iops,
431*4882a593Smuzhiyun 			    &bpf_iter_fops);
432*4882a593Smuzhiyun 	dput(dentry);
433*4882a593Smuzhiyun 	inode_unlock(parent->d_inode);
434*4882a593Smuzhiyun 	return ret;
435*4882a593Smuzhiyun }
436*4882a593Smuzhiyun 
bpf_obj_do_pin(const char __user * pathname,void * raw,enum bpf_type type)437*4882a593Smuzhiyun static int bpf_obj_do_pin(const char __user *pathname, void *raw,
438*4882a593Smuzhiyun 			  enum bpf_type type)
439*4882a593Smuzhiyun {
440*4882a593Smuzhiyun 	struct dentry *dentry;
441*4882a593Smuzhiyun 	struct inode *dir;
442*4882a593Smuzhiyun 	struct path path;
443*4882a593Smuzhiyun 	umode_t mode;
444*4882a593Smuzhiyun 	int ret;
445*4882a593Smuzhiyun 
446*4882a593Smuzhiyun 	dentry = user_path_create(AT_FDCWD, pathname, &path, 0);
447*4882a593Smuzhiyun 	if (IS_ERR(dentry))
448*4882a593Smuzhiyun 		return PTR_ERR(dentry);
449*4882a593Smuzhiyun 
450*4882a593Smuzhiyun 	mode = S_IFREG | ((S_IRUSR | S_IWUSR) & ~current_umask());
451*4882a593Smuzhiyun 
452*4882a593Smuzhiyun 	ret = security_path_mknod(&path, dentry, mode, 0);
453*4882a593Smuzhiyun 	if (ret)
454*4882a593Smuzhiyun 		goto out;
455*4882a593Smuzhiyun 
456*4882a593Smuzhiyun 	dir = d_inode(path.dentry);
457*4882a593Smuzhiyun 	if (dir->i_op != &bpf_dir_iops) {
458*4882a593Smuzhiyun 		ret = -EPERM;
459*4882a593Smuzhiyun 		goto out;
460*4882a593Smuzhiyun 	}
461*4882a593Smuzhiyun 
462*4882a593Smuzhiyun 	switch (type) {
463*4882a593Smuzhiyun 	case BPF_TYPE_PROG:
464*4882a593Smuzhiyun 		ret = vfs_mkobj(dentry, mode, bpf_mkprog, raw);
465*4882a593Smuzhiyun 		break;
466*4882a593Smuzhiyun 	case BPF_TYPE_MAP:
467*4882a593Smuzhiyun 		ret = vfs_mkobj(dentry, mode, bpf_mkmap, raw);
468*4882a593Smuzhiyun 		break;
469*4882a593Smuzhiyun 	case BPF_TYPE_LINK:
470*4882a593Smuzhiyun 		ret = vfs_mkobj(dentry, mode, bpf_mklink, raw);
471*4882a593Smuzhiyun 		break;
472*4882a593Smuzhiyun 	default:
473*4882a593Smuzhiyun 		ret = -EPERM;
474*4882a593Smuzhiyun 	}
475*4882a593Smuzhiyun out:
476*4882a593Smuzhiyun 	done_path_create(&path, dentry);
477*4882a593Smuzhiyun 	return ret;
478*4882a593Smuzhiyun }
479*4882a593Smuzhiyun 
bpf_obj_pin_user(u32 ufd,const char __user * pathname)480*4882a593Smuzhiyun int bpf_obj_pin_user(u32 ufd, const char __user *pathname)
481*4882a593Smuzhiyun {
482*4882a593Smuzhiyun 	enum bpf_type type;
483*4882a593Smuzhiyun 	void *raw;
484*4882a593Smuzhiyun 	int ret;
485*4882a593Smuzhiyun 
486*4882a593Smuzhiyun 	raw = bpf_fd_probe_obj(ufd, &type);
487*4882a593Smuzhiyun 	if (IS_ERR(raw))
488*4882a593Smuzhiyun 		return PTR_ERR(raw);
489*4882a593Smuzhiyun 
490*4882a593Smuzhiyun 	ret = bpf_obj_do_pin(pathname, raw, type);
491*4882a593Smuzhiyun 	if (ret != 0)
492*4882a593Smuzhiyun 		bpf_any_put(raw, type);
493*4882a593Smuzhiyun 
494*4882a593Smuzhiyun 	return ret;
495*4882a593Smuzhiyun }
496*4882a593Smuzhiyun 
bpf_obj_do_get(const char __user * pathname,enum bpf_type * type,int flags)497*4882a593Smuzhiyun static void *bpf_obj_do_get(const char __user *pathname,
498*4882a593Smuzhiyun 			    enum bpf_type *type, int flags)
499*4882a593Smuzhiyun {
500*4882a593Smuzhiyun 	struct inode *inode;
501*4882a593Smuzhiyun 	struct path path;
502*4882a593Smuzhiyun 	void *raw;
503*4882a593Smuzhiyun 	int ret;
504*4882a593Smuzhiyun 
505*4882a593Smuzhiyun 	ret = user_path_at(AT_FDCWD, pathname, LOOKUP_FOLLOW, &path);
506*4882a593Smuzhiyun 	if (ret)
507*4882a593Smuzhiyun 		return ERR_PTR(ret);
508*4882a593Smuzhiyun 
509*4882a593Smuzhiyun 	inode = d_backing_inode(path.dentry);
510*4882a593Smuzhiyun 	ret = inode_permission(inode, ACC_MODE(flags));
511*4882a593Smuzhiyun 	if (ret)
512*4882a593Smuzhiyun 		goto out;
513*4882a593Smuzhiyun 
514*4882a593Smuzhiyun 	ret = bpf_inode_type(inode, type);
515*4882a593Smuzhiyun 	if (ret)
516*4882a593Smuzhiyun 		goto out;
517*4882a593Smuzhiyun 
518*4882a593Smuzhiyun 	raw = bpf_any_get(inode->i_private, *type);
519*4882a593Smuzhiyun 	if (!IS_ERR(raw))
520*4882a593Smuzhiyun 		touch_atime(&path);
521*4882a593Smuzhiyun 
522*4882a593Smuzhiyun 	path_put(&path);
523*4882a593Smuzhiyun 	return raw;
524*4882a593Smuzhiyun out:
525*4882a593Smuzhiyun 	path_put(&path);
526*4882a593Smuzhiyun 	return ERR_PTR(ret);
527*4882a593Smuzhiyun }
528*4882a593Smuzhiyun 
bpf_obj_get_user(const char __user * pathname,int flags)529*4882a593Smuzhiyun int bpf_obj_get_user(const char __user *pathname, int flags)
530*4882a593Smuzhiyun {
531*4882a593Smuzhiyun 	enum bpf_type type = BPF_TYPE_UNSPEC;
532*4882a593Smuzhiyun 	int f_flags;
533*4882a593Smuzhiyun 	void *raw;
534*4882a593Smuzhiyun 	int ret;
535*4882a593Smuzhiyun 
536*4882a593Smuzhiyun 	f_flags = bpf_get_file_flag(flags);
537*4882a593Smuzhiyun 	if (f_flags < 0)
538*4882a593Smuzhiyun 		return f_flags;
539*4882a593Smuzhiyun 
540*4882a593Smuzhiyun 	raw = bpf_obj_do_get(pathname, &type, f_flags);
541*4882a593Smuzhiyun 	if (IS_ERR(raw))
542*4882a593Smuzhiyun 		return PTR_ERR(raw);
543*4882a593Smuzhiyun 
544*4882a593Smuzhiyun 	if (type == BPF_TYPE_PROG)
545*4882a593Smuzhiyun 		ret = bpf_prog_new_fd(raw);
546*4882a593Smuzhiyun 	else if (type == BPF_TYPE_MAP)
547*4882a593Smuzhiyun 		ret = bpf_map_new_fd(raw, f_flags);
548*4882a593Smuzhiyun 	else if (type == BPF_TYPE_LINK)
549*4882a593Smuzhiyun 		ret = (f_flags != O_RDWR) ? -EINVAL : bpf_link_new_fd(raw);
550*4882a593Smuzhiyun 	else
551*4882a593Smuzhiyun 		return -ENOENT;
552*4882a593Smuzhiyun 
553*4882a593Smuzhiyun 	if (ret < 0)
554*4882a593Smuzhiyun 		bpf_any_put(raw, type);
555*4882a593Smuzhiyun 	return ret;
556*4882a593Smuzhiyun }
557*4882a593Smuzhiyun 
__get_prog_inode(struct inode * inode,enum bpf_prog_type type)558*4882a593Smuzhiyun static struct bpf_prog *__get_prog_inode(struct inode *inode, enum bpf_prog_type type)
559*4882a593Smuzhiyun {
560*4882a593Smuzhiyun 	struct bpf_prog *prog;
561*4882a593Smuzhiyun 	int ret = inode_permission(inode, MAY_READ);
562*4882a593Smuzhiyun 	if (ret)
563*4882a593Smuzhiyun 		return ERR_PTR(ret);
564*4882a593Smuzhiyun 
565*4882a593Smuzhiyun 	if (inode->i_op == &bpf_map_iops)
566*4882a593Smuzhiyun 		return ERR_PTR(-EINVAL);
567*4882a593Smuzhiyun 	if (inode->i_op == &bpf_link_iops)
568*4882a593Smuzhiyun 		return ERR_PTR(-EINVAL);
569*4882a593Smuzhiyun 	if (inode->i_op != &bpf_prog_iops)
570*4882a593Smuzhiyun 		return ERR_PTR(-EACCES);
571*4882a593Smuzhiyun 
572*4882a593Smuzhiyun 	prog = inode->i_private;
573*4882a593Smuzhiyun 
574*4882a593Smuzhiyun 	ret = security_bpf_prog(prog);
575*4882a593Smuzhiyun 	if (ret < 0)
576*4882a593Smuzhiyun 		return ERR_PTR(ret);
577*4882a593Smuzhiyun 
578*4882a593Smuzhiyun 	if (!bpf_prog_get_ok(prog, &type, false))
579*4882a593Smuzhiyun 		return ERR_PTR(-EINVAL);
580*4882a593Smuzhiyun 
581*4882a593Smuzhiyun 	bpf_prog_inc(prog);
582*4882a593Smuzhiyun 	return prog;
583*4882a593Smuzhiyun }
584*4882a593Smuzhiyun 
bpf_prog_get_type_path(const char * name,enum bpf_prog_type type)585*4882a593Smuzhiyun struct bpf_prog *bpf_prog_get_type_path(const char *name, enum bpf_prog_type type)
586*4882a593Smuzhiyun {
587*4882a593Smuzhiyun 	struct bpf_prog *prog;
588*4882a593Smuzhiyun 	struct path path;
589*4882a593Smuzhiyun 	int ret = kern_path(name, LOOKUP_FOLLOW, &path);
590*4882a593Smuzhiyun 	if (ret)
591*4882a593Smuzhiyun 		return ERR_PTR(ret);
592*4882a593Smuzhiyun 	prog = __get_prog_inode(d_backing_inode(path.dentry), type);
593*4882a593Smuzhiyun 	if (!IS_ERR(prog))
594*4882a593Smuzhiyun 		touch_atime(&path);
595*4882a593Smuzhiyun 	path_put(&path);
596*4882a593Smuzhiyun 	return prog;
597*4882a593Smuzhiyun }
598*4882a593Smuzhiyun EXPORT_SYMBOL(bpf_prog_get_type_path);
599*4882a593Smuzhiyun 
600*4882a593Smuzhiyun /*
601*4882a593Smuzhiyun  * Display the mount options in /proc/mounts.
602*4882a593Smuzhiyun  */
bpf_show_options(struct seq_file * m,struct dentry * root)603*4882a593Smuzhiyun static int bpf_show_options(struct seq_file *m, struct dentry *root)
604*4882a593Smuzhiyun {
605*4882a593Smuzhiyun 	umode_t mode = d_inode(root)->i_mode & S_IALLUGO & ~S_ISVTX;
606*4882a593Smuzhiyun 
607*4882a593Smuzhiyun 	if (mode != S_IRWXUGO)
608*4882a593Smuzhiyun 		seq_printf(m, ",mode=%o", mode);
609*4882a593Smuzhiyun 	return 0;
610*4882a593Smuzhiyun }
611*4882a593Smuzhiyun 
bpf_free_inode(struct inode * inode)612*4882a593Smuzhiyun static void bpf_free_inode(struct inode *inode)
613*4882a593Smuzhiyun {
614*4882a593Smuzhiyun 	enum bpf_type type;
615*4882a593Smuzhiyun 
616*4882a593Smuzhiyun 	if (S_ISLNK(inode->i_mode))
617*4882a593Smuzhiyun 		kfree(inode->i_link);
618*4882a593Smuzhiyun 	if (!bpf_inode_type(inode, &type))
619*4882a593Smuzhiyun 		bpf_any_put(inode->i_private, type);
620*4882a593Smuzhiyun 	free_inode_nonrcu(inode);
621*4882a593Smuzhiyun }
622*4882a593Smuzhiyun 
623*4882a593Smuzhiyun static const struct super_operations bpf_super_ops = {
624*4882a593Smuzhiyun 	.statfs		= simple_statfs,
625*4882a593Smuzhiyun 	.drop_inode	= generic_delete_inode,
626*4882a593Smuzhiyun 	.show_options	= bpf_show_options,
627*4882a593Smuzhiyun 	.free_inode	= bpf_free_inode,
628*4882a593Smuzhiyun };
629*4882a593Smuzhiyun 
630*4882a593Smuzhiyun enum {
631*4882a593Smuzhiyun 	OPT_MODE,
632*4882a593Smuzhiyun };
633*4882a593Smuzhiyun 
634*4882a593Smuzhiyun static const struct fs_parameter_spec bpf_fs_parameters[] = {
635*4882a593Smuzhiyun 	fsparam_u32oct	("mode",			OPT_MODE),
636*4882a593Smuzhiyun 	{}
637*4882a593Smuzhiyun };
638*4882a593Smuzhiyun 
639*4882a593Smuzhiyun struct bpf_mount_opts {
640*4882a593Smuzhiyun 	umode_t mode;
641*4882a593Smuzhiyun };
642*4882a593Smuzhiyun 
bpf_parse_param(struct fs_context * fc,struct fs_parameter * param)643*4882a593Smuzhiyun static int bpf_parse_param(struct fs_context *fc, struct fs_parameter *param)
644*4882a593Smuzhiyun {
645*4882a593Smuzhiyun 	struct bpf_mount_opts *opts = fc->fs_private;
646*4882a593Smuzhiyun 	struct fs_parse_result result;
647*4882a593Smuzhiyun 	int opt;
648*4882a593Smuzhiyun 
649*4882a593Smuzhiyun 	opt = fs_parse(fc, bpf_fs_parameters, param, &result);
650*4882a593Smuzhiyun 	if (opt < 0)
651*4882a593Smuzhiyun 		/* We might like to report bad mount options here, but
652*4882a593Smuzhiyun 		 * traditionally we've ignored all mount options, so we'd
653*4882a593Smuzhiyun 		 * better continue to ignore non-existing options for bpf.
654*4882a593Smuzhiyun 		 */
655*4882a593Smuzhiyun 		return opt == -ENOPARAM ? 0 : opt;
656*4882a593Smuzhiyun 
657*4882a593Smuzhiyun 	switch (opt) {
658*4882a593Smuzhiyun 	case OPT_MODE:
659*4882a593Smuzhiyun 		opts->mode = result.uint_32 & S_IALLUGO;
660*4882a593Smuzhiyun 		break;
661*4882a593Smuzhiyun 	}
662*4882a593Smuzhiyun 
663*4882a593Smuzhiyun 	return 0;
664*4882a593Smuzhiyun }
665*4882a593Smuzhiyun 
666*4882a593Smuzhiyun struct bpf_preload_ops *bpf_preload_ops;
667*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(bpf_preload_ops);
668*4882a593Smuzhiyun 
bpf_preload_mod_get(void)669*4882a593Smuzhiyun static bool bpf_preload_mod_get(void)
670*4882a593Smuzhiyun {
671*4882a593Smuzhiyun 	/* If bpf_preload.ko wasn't loaded earlier then load it now.
672*4882a593Smuzhiyun 	 * When bpf_preload is built into vmlinux the module's __init
673*4882a593Smuzhiyun 	 * function will populate it.
674*4882a593Smuzhiyun 	 */
675*4882a593Smuzhiyun 	if (!bpf_preload_ops) {
676*4882a593Smuzhiyun 		request_module("bpf_preload");
677*4882a593Smuzhiyun 		if (!bpf_preload_ops)
678*4882a593Smuzhiyun 			return false;
679*4882a593Smuzhiyun 	}
680*4882a593Smuzhiyun 	/* And grab the reference, so the module doesn't disappear while the
681*4882a593Smuzhiyun 	 * kernel is interacting with the kernel module and its UMD.
682*4882a593Smuzhiyun 	 */
683*4882a593Smuzhiyun 	if (!try_module_get(bpf_preload_ops->owner)) {
684*4882a593Smuzhiyun 		pr_err("bpf_preload module get failed.\n");
685*4882a593Smuzhiyun 		return false;
686*4882a593Smuzhiyun 	}
687*4882a593Smuzhiyun 	return true;
688*4882a593Smuzhiyun }
689*4882a593Smuzhiyun 
bpf_preload_mod_put(void)690*4882a593Smuzhiyun static void bpf_preload_mod_put(void)
691*4882a593Smuzhiyun {
692*4882a593Smuzhiyun 	if (bpf_preload_ops)
693*4882a593Smuzhiyun 		/* now user can "rmmod bpf_preload" if necessary */
694*4882a593Smuzhiyun 		module_put(bpf_preload_ops->owner);
695*4882a593Smuzhiyun }
696*4882a593Smuzhiyun 
697*4882a593Smuzhiyun static DEFINE_MUTEX(bpf_preload_lock);
698*4882a593Smuzhiyun 
populate_bpffs(struct dentry * parent)699*4882a593Smuzhiyun static int populate_bpffs(struct dentry *parent)
700*4882a593Smuzhiyun {
701*4882a593Smuzhiyun 	struct bpf_preload_info objs[BPF_PRELOAD_LINKS] = {};
702*4882a593Smuzhiyun 	struct bpf_link *links[BPF_PRELOAD_LINKS] = {};
703*4882a593Smuzhiyun 	int err = 0, i;
704*4882a593Smuzhiyun 
705*4882a593Smuzhiyun 	/* grab the mutex to make sure the kernel interactions with bpf_preload
706*4882a593Smuzhiyun 	 * UMD are serialized
707*4882a593Smuzhiyun 	 */
708*4882a593Smuzhiyun 	mutex_lock(&bpf_preload_lock);
709*4882a593Smuzhiyun 
710*4882a593Smuzhiyun 	/* if bpf_preload.ko wasn't built into vmlinux then load it */
711*4882a593Smuzhiyun 	if (!bpf_preload_mod_get())
712*4882a593Smuzhiyun 		goto out;
713*4882a593Smuzhiyun 
714*4882a593Smuzhiyun 	if (!bpf_preload_ops->info.tgid) {
715*4882a593Smuzhiyun 		/* preload() will start UMD that will load BPF iterator programs */
716*4882a593Smuzhiyun 		err = bpf_preload_ops->preload(objs);
717*4882a593Smuzhiyun 		if (err)
718*4882a593Smuzhiyun 			goto out_put;
719*4882a593Smuzhiyun 		for (i = 0; i < BPF_PRELOAD_LINKS; i++) {
720*4882a593Smuzhiyun 			links[i] = bpf_link_by_id(objs[i].link_id);
721*4882a593Smuzhiyun 			if (IS_ERR(links[i])) {
722*4882a593Smuzhiyun 				err = PTR_ERR(links[i]);
723*4882a593Smuzhiyun 				goto out_put;
724*4882a593Smuzhiyun 			}
725*4882a593Smuzhiyun 		}
726*4882a593Smuzhiyun 		for (i = 0; i < BPF_PRELOAD_LINKS; i++) {
727*4882a593Smuzhiyun 			err = bpf_iter_link_pin_kernel(parent,
728*4882a593Smuzhiyun 						       objs[i].link_name, links[i]);
729*4882a593Smuzhiyun 			if (err)
730*4882a593Smuzhiyun 				goto out_put;
731*4882a593Smuzhiyun 			/* do not unlink successfully pinned links even
732*4882a593Smuzhiyun 			 * if later link fails to pin
733*4882a593Smuzhiyun 			 */
734*4882a593Smuzhiyun 			links[i] = NULL;
735*4882a593Smuzhiyun 		}
736*4882a593Smuzhiyun 		/* finish() will tell UMD process to exit */
737*4882a593Smuzhiyun 		err = bpf_preload_ops->finish();
738*4882a593Smuzhiyun 		if (err)
739*4882a593Smuzhiyun 			goto out_put;
740*4882a593Smuzhiyun 	}
741*4882a593Smuzhiyun out_put:
742*4882a593Smuzhiyun 	bpf_preload_mod_put();
743*4882a593Smuzhiyun out:
744*4882a593Smuzhiyun 	mutex_unlock(&bpf_preload_lock);
745*4882a593Smuzhiyun 	for (i = 0; i < BPF_PRELOAD_LINKS && err; i++)
746*4882a593Smuzhiyun 		if (!IS_ERR_OR_NULL(links[i]))
747*4882a593Smuzhiyun 			bpf_link_put(links[i]);
748*4882a593Smuzhiyun 	return err;
749*4882a593Smuzhiyun }
750*4882a593Smuzhiyun 
bpf_fill_super(struct super_block * sb,struct fs_context * fc)751*4882a593Smuzhiyun static int bpf_fill_super(struct super_block *sb, struct fs_context *fc)
752*4882a593Smuzhiyun {
753*4882a593Smuzhiyun 	static const struct tree_descr bpf_rfiles[] = { { "" } };
754*4882a593Smuzhiyun 	struct bpf_mount_opts *opts = fc->fs_private;
755*4882a593Smuzhiyun 	struct inode *inode;
756*4882a593Smuzhiyun 	int ret;
757*4882a593Smuzhiyun 
758*4882a593Smuzhiyun 	ret = simple_fill_super(sb, BPF_FS_MAGIC, bpf_rfiles);
759*4882a593Smuzhiyun 	if (ret)
760*4882a593Smuzhiyun 		return ret;
761*4882a593Smuzhiyun 
762*4882a593Smuzhiyun 	sb->s_op = &bpf_super_ops;
763*4882a593Smuzhiyun 
764*4882a593Smuzhiyun 	inode = sb->s_root->d_inode;
765*4882a593Smuzhiyun 	inode->i_op = &bpf_dir_iops;
766*4882a593Smuzhiyun 	inode->i_mode &= ~S_IALLUGO;
767*4882a593Smuzhiyun 	populate_bpffs(sb->s_root);
768*4882a593Smuzhiyun 	inode->i_mode |= S_ISVTX | opts->mode;
769*4882a593Smuzhiyun 	return 0;
770*4882a593Smuzhiyun }
771*4882a593Smuzhiyun 
bpf_get_tree(struct fs_context * fc)772*4882a593Smuzhiyun static int bpf_get_tree(struct fs_context *fc)
773*4882a593Smuzhiyun {
774*4882a593Smuzhiyun 	return get_tree_nodev(fc, bpf_fill_super);
775*4882a593Smuzhiyun }
776*4882a593Smuzhiyun 
bpf_free_fc(struct fs_context * fc)777*4882a593Smuzhiyun static void bpf_free_fc(struct fs_context *fc)
778*4882a593Smuzhiyun {
779*4882a593Smuzhiyun 	kfree(fc->fs_private);
780*4882a593Smuzhiyun }
781*4882a593Smuzhiyun 
782*4882a593Smuzhiyun static const struct fs_context_operations bpf_context_ops = {
783*4882a593Smuzhiyun 	.free		= bpf_free_fc,
784*4882a593Smuzhiyun 	.parse_param	= bpf_parse_param,
785*4882a593Smuzhiyun 	.get_tree	= bpf_get_tree,
786*4882a593Smuzhiyun };
787*4882a593Smuzhiyun 
788*4882a593Smuzhiyun /*
789*4882a593Smuzhiyun  * Set up the filesystem mount context.
790*4882a593Smuzhiyun  */
bpf_init_fs_context(struct fs_context * fc)791*4882a593Smuzhiyun static int bpf_init_fs_context(struct fs_context *fc)
792*4882a593Smuzhiyun {
793*4882a593Smuzhiyun 	struct bpf_mount_opts *opts;
794*4882a593Smuzhiyun 
795*4882a593Smuzhiyun 	opts = kzalloc(sizeof(struct bpf_mount_opts), GFP_KERNEL);
796*4882a593Smuzhiyun 	if (!opts)
797*4882a593Smuzhiyun 		return -ENOMEM;
798*4882a593Smuzhiyun 
799*4882a593Smuzhiyun 	opts->mode = S_IRWXUGO;
800*4882a593Smuzhiyun 
801*4882a593Smuzhiyun 	fc->fs_private = opts;
802*4882a593Smuzhiyun 	fc->ops = &bpf_context_ops;
803*4882a593Smuzhiyun 	return 0;
804*4882a593Smuzhiyun }
805*4882a593Smuzhiyun 
806*4882a593Smuzhiyun static struct file_system_type bpf_fs_type = {
807*4882a593Smuzhiyun 	.owner		= THIS_MODULE,
808*4882a593Smuzhiyun 	.name		= "bpf",
809*4882a593Smuzhiyun 	.init_fs_context = bpf_init_fs_context,
810*4882a593Smuzhiyun 	.parameters	= bpf_fs_parameters,
811*4882a593Smuzhiyun 	.kill_sb	= kill_litter_super,
812*4882a593Smuzhiyun };
813*4882a593Smuzhiyun 
bpf_init(void)814*4882a593Smuzhiyun static int __init bpf_init(void)
815*4882a593Smuzhiyun {
816*4882a593Smuzhiyun 	int ret;
817*4882a593Smuzhiyun 
818*4882a593Smuzhiyun 	mutex_init(&bpf_preload_lock);
819*4882a593Smuzhiyun 
820*4882a593Smuzhiyun 	ret = sysfs_create_mount_point(fs_kobj, "bpf");
821*4882a593Smuzhiyun 	if (ret)
822*4882a593Smuzhiyun 		return ret;
823*4882a593Smuzhiyun 
824*4882a593Smuzhiyun 	ret = register_filesystem(&bpf_fs_type);
825*4882a593Smuzhiyun 	if (ret)
826*4882a593Smuzhiyun 		sysfs_remove_mount_point(fs_kobj, "bpf");
827*4882a593Smuzhiyun 
828*4882a593Smuzhiyun 	return ret;
829*4882a593Smuzhiyun }
830*4882a593Smuzhiyun fs_initcall(bpf_init);
831