xref: /openbmc/linux/tools/bpf/bpftool/common.c (revision a48acad7)
1 // SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause)
2 /* Copyright (C) 2017-2018 Netronome Systems, Inc. */
3 
4 #ifndef _GNU_SOURCE
5 #define _GNU_SOURCE
6 #endif
7 #include <ctype.h>
8 #include <errno.h>
9 #include <fcntl.h>
10 #include <ftw.h>
11 #include <libgen.h>
12 #include <mntent.h>
13 #include <stdbool.h>
14 #include <stdio.h>
15 #include <stdlib.h>
16 #include <string.h>
17 #include <unistd.h>
18 #include <net/if.h>
19 #include <sys/mount.h>
20 #include <sys/resource.h>
21 #include <sys/stat.h>
22 #include <sys/vfs.h>
23 
24 #include <linux/filter.h>
25 #include <linux/limits.h>
26 #include <linux/magic.h>
27 #include <linux/unistd.h>
28 
29 #include <bpf/bpf.h>
30 #include <bpf/hashmap.h>
31 #include <bpf/libbpf.h> /* libbpf_num_possible_cpus */
32 #include <bpf/btf.h>
33 
34 #include "main.h"
35 
36 #ifndef BPF_FS_MAGIC
37 #define BPF_FS_MAGIC		0xcafe4a11
38 #endif
39 
40 void p_err(const char *fmt, ...)
41 {
42 	va_list ap;
43 
44 	va_start(ap, fmt);
45 	if (json_output) {
46 		jsonw_start_object(json_wtr);
47 		jsonw_name(json_wtr, "error");
48 		jsonw_vprintf_enquote(json_wtr, fmt, ap);
49 		jsonw_end_object(json_wtr);
50 	} else {
51 		fprintf(stderr, "Error: ");
52 		vfprintf(stderr, fmt, ap);
53 		fprintf(stderr, "\n");
54 	}
55 	va_end(ap);
56 }
57 
58 void p_info(const char *fmt, ...)
59 {
60 	va_list ap;
61 
62 	if (json_output)
63 		return;
64 
65 	va_start(ap, fmt);
66 	vfprintf(stderr, fmt, ap);
67 	fprintf(stderr, "\n");
68 	va_end(ap);
69 }
70 
71 static bool is_bpffs(char *path)
72 {
73 	struct statfs st_fs;
74 
75 	if (statfs(path, &st_fs) < 0)
76 		return false;
77 
78 	return (unsigned long)st_fs.f_type == BPF_FS_MAGIC;
79 }
80 
81 /* Probe whether kernel switched from memlock-based (RLIMIT_MEMLOCK) to
82  * memcg-based memory accounting for BPF maps and programs. This was done in
83  * commit 97306be45fbe ("Merge branch 'switch to memcg-based memory
84  * accounting'"), in Linux 5.11.
85  *
86  * Libbpf also offers to probe for memcg-based accounting vs rlimit, but does
87  * so by checking for the availability of a given BPF helper and this has
88  * failed on some kernels with backports in the past, see commit 6b4384ff1088
89  * ("Revert "bpftool: Use libbpf 1.0 API mode instead of RLIMIT_MEMLOCK"").
90  * Instead, we can probe by lowering the process-based rlimit to 0, trying to
91  * load a BPF object, and resetting the rlimit. If the load succeeds then
92  * memcg-based accounting is supported.
93  *
94  * This would be too dangerous to do in the library, because multithreaded
95  * applications might attempt to load items while the rlimit is at 0. Given
96  * that bpftool is single-threaded, this is fine to do here.
97  */
98 static bool known_to_need_rlimit(void)
99 {
100 	struct rlimit rlim_init, rlim_cur_zero = {};
101 	struct bpf_insn insns[] = {
102 		BPF_MOV64_IMM(BPF_REG_0, 0),
103 		BPF_EXIT_INSN(),
104 	};
105 	size_t insn_cnt = ARRAY_SIZE(insns);
106 	union bpf_attr attr;
107 	int prog_fd, err;
108 
109 	memset(&attr, 0, sizeof(attr));
110 	attr.prog_type = BPF_PROG_TYPE_SOCKET_FILTER;
111 	attr.insns = ptr_to_u64(insns);
112 	attr.insn_cnt = insn_cnt;
113 	attr.license = ptr_to_u64("GPL");
114 
115 	if (getrlimit(RLIMIT_MEMLOCK, &rlim_init))
116 		return false;
117 
118 	/* Drop the soft limit to zero. We maintain the hard limit to its
119 	 * current value, because lowering it would be a permanent operation
120 	 * for unprivileged users.
121 	 */
122 	rlim_cur_zero.rlim_max = rlim_init.rlim_max;
123 	if (setrlimit(RLIMIT_MEMLOCK, &rlim_cur_zero))
124 		return false;
125 
126 	/* Do not use bpf_prog_load() from libbpf here, because it calls
127 	 * bump_rlimit_memlock(), interfering with the current probe.
128 	 */
129 	prog_fd = syscall(__NR_bpf, BPF_PROG_LOAD, &attr, sizeof(attr));
130 	err = errno;
131 
132 	/* reset soft rlimit to its initial value */
133 	setrlimit(RLIMIT_MEMLOCK, &rlim_init);
134 
135 	if (prog_fd < 0)
136 		return err == EPERM;
137 
138 	close(prog_fd);
139 	return false;
140 }
141 
142 void set_max_rlimit(void)
143 {
144 	struct rlimit rinf = { RLIM_INFINITY, RLIM_INFINITY };
145 
146 	if (known_to_need_rlimit())
147 		setrlimit(RLIMIT_MEMLOCK, &rinf);
148 }
149 
150 static int
151 mnt_fs(const char *target, const char *type, char *buff, size_t bufflen)
152 {
153 	bool bind_done = false;
154 
155 	while (mount("", target, "none", MS_PRIVATE | MS_REC, NULL)) {
156 		if (errno != EINVAL || bind_done) {
157 			snprintf(buff, bufflen,
158 				 "mount --make-private %s failed: %s",
159 				 target, strerror(errno));
160 			return -1;
161 		}
162 
163 		if (mount(target, target, "none", MS_BIND, NULL)) {
164 			snprintf(buff, bufflen,
165 				 "mount --bind %s %s failed: %s",
166 				 target, target, strerror(errno));
167 			return -1;
168 		}
169 
170 		bind_done = true;
171 	}
172 
173 	if (mount(type, target, type, 0, "mode=0700")) {
174 		snprintf(buff, bufflen, "mount -t %s %s %s failed: %s",
175 			 type, type, target, strerror(errno));
176 		return -1;
177 	}
178 
179 	return 0;
180 }
181 
182 int mount_tracefs(const char *target)
183 {
184 	char err_str[ERR_MAX_LEN];
185 	int err;
186 
187 	err = mnt_fs(target, "tracefs", err_str, ERR_MAX_LEN);
188 	if (err) {
189 		err_str[ERR_MAX_LEN - 1] = '\0';
190 		p_err("can't mount tracefs: %s", err_str);
191 	}
192 
193 	return err;
194 }
195 
196 int open_obj_pinned(const char *path, bool quiet)
197 {
198 	char *pname;
199 	int fd = -1;
200 
201 	pname = strdup(path);
202 	if (!pname) {
203 		if (!quiet)
204 			p_err("mem alloc failed");
205 		goto out_ret;
206 	}
207 
208 	fd = bpf_obj_get(pname);
209 	if (fd < 0) {
210 		if (!quiet)
211 			p_err("bpf obj get (%s): %s", pname,
212 			      errno == EACCES && !is_bpffs(dirname(pname)) ?
213 			    "directory not in bpf file system (bpffs)" :
214 			    strerror(errno));
215 		goto out_free;
216 	}
217 
218 out_free:
219 	free(pname);
220 out_ret:
221 	return fd;
222 }
223 
224 int open_obj_pinned_any(const char *path, enum bpf_obj_type exp_type)
225 {
226 	enum bpf_obj_type type;
227 	int fd;
228 
229 	fd = open_obj_pinned(path, false);
230 	if (fd < 0)
231 		return -1;
232 
233 	type = get_fd_type(fd);
234 	if (type < 0) {
235 		close(fd);
236 		return type;
237 	}
238 	if (type != exp_type) {
239 		p_err("incorrect object type: %s", get_fd_type_name(type));
240 		close(fd);
241 		return -1;
242 	}
243 
244 	return fd;
245 }
246 
247 int mount_bpffs_for_pin(const char *name)
248 {
249 	char err_str[ERR_MAX_LEN];
250 	char *file;
251 	char *dir;
252 	int err = 0;
253 
254 	file = malloc(strlen(name) + 1);
255 	if (!file) {
256 		p_err("mem alloc failed");
257 		return -1;
258 	}
259 
260 	strcpy(file, name);
261 	dir = dirname(file);
262 
263 	if (is_bpffs(dir))
264 		/* nothing to do if already mounted */
265 		goto out_free;
266 
267 	if (block_mount) {
268 		p_err("no BPF file system found, not mounting it due to --nomount option");
269 		err = -1;
270 		goto out_free;
271 	}
272 
273 	err = mnt_fs(dir, "bpf", err_str, ERR_MAX_LEN);
274 	if (err) {
275 		err_str[ERR_MAX_LEN - 1] = '\0';
276 		p_err("can't mount BPF file system to pin the object (%s): %s",
277 		      name, err_str);
278 	}
279 
280 out_free:
281 	free(file);
282 	return err;
283 }
284 
285 int do_pin_fd(int fd, const char *name)
286 {
287 	int err;
288 
289 	err = mount_bpffs_for_pin(name);
290 	if (err)
291 		return err;
292 
293 	err = bpf_obj_pin(fd, name);
294 	if (err)
295 		p_err("can't pin the object (%s): %s", name, strerror(errno));
296 
297 	return err;
298 }
299 
300 int do_pin_any(int argc, char **argv, int (*get_fd)(int *, char ***))
301 {
302 	int err;
303 	int fd;
304 
305 	if (!REQ_ARGS(3))
306 		return -EINVAL;
307 
308 	fd = get_fd(&argc, &argv);
309 	if (fd < 0)
310 		return fd;
311 
312 	err = do_pin_fd(fd, *argv);
313 
314 	close(fd);
315 	return err;
316 }
317 
318 const char *get_fd_type_name(enum bpf_obj_type type)
319 {
320 	static const char * const names[] = {
321 		[BPF_OBJ_UNKNOWN]	= "unknown",
322 		[BPF_OBJ_PROG]		= "prog",
323 		[BPF_OBJ_MAP]		= "map",
324 		[BPF_OBJ_LINK]		= "link",
325 	};
326 
327 	if (type < 0 || type >= ARRAY_SIZE(names) || !names[type])
328 		return names[BPF_OBJ_UNKNOWN];
329 
330 	return names[type];
331 }
332 
333 void get_prog_full_name(const struct bpf_prog_info *prog_info, int prog_fd,
334 			char *name_buff, size_t buff_len)
335 {
336 	const char *prog_name = prog_info->name;
337 	const struct btf_type *func_type;
338 	const struct bpf_func_info finfo = {};
339 	struct bpf_prog_info info = {};
340 	__u32 info_len = sizeof(info);
341 	struct btf *prog_btf = NULL;
342 
343 	if (buff_len <= BPF_OBJ_NAME_LEN ||
344 	    strlen(prog_info->name) < BPF_OBJ_NAME_LEN - 1)
345 		goto copy_name;
346 
347 	if (!prog_info->btf_id || prog_info->nr_func_info == 0)
348 		goto copy_name;
349 
350 	info.nr_func_info = 1;
351 	info.func_info_rec_size = prog_info->func_info_rec_size;
352 	if (info.func_info_rec_size > sizeof(finfo))
353 		info.func_info_rec_size = sizeof(finfo);
354 	info.func_info = ptr_to_u64(&finfo);
355 
356 	if (bpf_obj_get_info_by_fd(prog_fd, &info, &info_len))
357 		goto copy_name;
358 
359 	prog_btf = btf__load_from_kernel_by_id(info.btf_id);
360 	if (!prog_btf)
361 		goto copy_name;
362 
363 	func_type = btf__type_by_id(prog_btf, finfo.type_id);
364 	if (!func_type || !btf_is_func(func_type))
365 		goto copy_name;
366 
367 	prog_name = btf__name_by_offset(prog_btf, func_type->name_off);
368 
369 copy_name:
370 	snprintf(name_buff, buff_len, "%s", prog_name);
371 
372 	if (prog_btf)
373 		btf__free(prog_btf);
374 }
375 
376 int get_fd_type(int fd)
377 {
378 	char path[PATH_MAX];
379 	char buf[512];
380 	ssize_t n;
381 
382 	snprintf(path, sizeof(path), "/proc/self/fd/%d", fd);
383 
384 	n = readlink(path, buf, sizeof(buf));
385 	if (n < 0) {
386 		p_err("can't read link type: %s", strerror(errno));
387 		return -1;
388 	}
389 	if (n == sizeof(path)) {
390 		p_err("can't read link type: path too long!");
391 		return -1;
392 	}
393 
394 	if (strstr(buf, "bpf-map"))
395 		return BPF_OBJ_MAP;
396 	else if (strstr(buf, "bpf-prog"))
397 		return BPF_OBJ_PROG;
398 	else if (strstr(buf, "bpf-link"))
399 		return BPF_OBJ_LINK;
400 
401 	return BPF_OBJ_UNKNOWN;
402 }
403 
404 char *get_fdinfo(int fd, const char *key)
405 {
406 	char path[PATH_MAX];
407 	char *line = NULL;
408 	size_t line_n = 0;
409 	ssize_t n;
410 	FILE *fdi;
411 
412 	snprintf(path, sizeof(path), "/proc/self/fdinfo/%d", fd);
413 
414 	fdi = fopen(path, "r");
415 	if (!fdi)
416 		return NULL;
417 
418 	while ((n = getline(&line, &line_n, fdi)) > 0) {
419 		char *value;
420 		int len;
421 
422 		if (!strstr(line, key))
423 			continue;
424 
425 		fclose(fdi);
426 
427 		value = strchr(line, '\t');
428 		if (!value || !value[1]) {
429 			free(line);
430 			return NULL;
431 		}
432 		value++;
433 
434 		len = strlen(value);
435 		memmove(line, value, len);
436 		line[len - 1] = '\0';
437 
438 		return line;
439 	}
440 
441 	free(line);
442 	fclose(fdi);
443 	return NULL;
444 }
445 
446 void print_data_json(uint8_t *data, size_t len)
447 {
448 	unsigned int i;
449 
450 	jsonw_start_array(json_wtr);
451 	for (i = 0; i < len; i++)
452 		jsonw_printf(json_wtr, "%d", data[i]);
453 	jsonw_end_array(json_wtr);
454 }
455 
456 void print_hex_data_json(uint8_t *data, size_t len)
457 {
458 	unsigned int i;
459 
460 	jsonw_start_array(json_wtr);
461 	for (i = 0; i < len; i++)
462 		jsonw_printf(json_wtr, "\"0x%02hhx\"", data[i]);
463 	jsonw_end_array(json_wtr);
464 }
465 
466 /* extra params for nftw cb */
467 static struct hashmap *build_fn_table;
468 static enum bpf_obj_type build_fn_type;
469 
470 static int do_build_table_cb(const char *fpath, const struct stat *sb,
471 			     int typeflag, struct FTW *ftwbuf)
472 {
473 	struct bpf_prog_info pinned_info;
474 	__u32 len = sizeof(pinned_info);
475 	enum bpf_obj_type objtype;
476 	int fd, err = 0;
477 	char *path;
478 
479 	if (typeflag != FTW_F)
480 		goto out_ret;
481 
482 	fd = open_obj_pinned(fpath, true);
483 	if (fd < 0)
484 		goto out_ret;
485 
486 	objtype = get_fd_type(fd);
487 	if (objtype != build_fn_type)
488 		goto out_close;
489 
490 	memset(&pinned_info, 0, sizeof(pinned_info));
491 	if (bpf_obj_get_info_by_fd(fd, &pinned_info, &len))
492 		goto out_close;
493 
494 	path = strdup(fpath);
495 	if (!path) {
496 		err = -1;
497 		goto out_close;
498 	}
499 
500 	err = hashmap__append(build_fn_table, pinned_info.id, path);
501 	if (err) {
502 		p_err("failed to append entry to hashmap for ID %u, path '%s': %s",
503 		      pinned_info.id, path, strerror(errno));
504 		goto out_close;
505 	}
506 
507 out_close:
508 	close(fd);
509 out_ret:
510 	return err;
511 }
512 
513 int build_pinned_obj_table(struct hashmap *tab,
514 			   enum bpf_obj_type type)
515 {
516 	struct mntent *mntent = NULL;
517 	FILE *mntfile = NULL;
518 	int flags = FTW_PHYS;
519 	int nopenfd = 16;
520 	int err = 0;
521 
522 	mntfile = setmntent("/proc/mounts", "r");
523 	if (!mntfile)
524 		return -1;
525 
526 	build_fn_table = tab;
527 	build_fn_type = type;
528 
529 	while ((mntent = getmntent(mntfile))) {
530 		char *path = mntent->mnt_dir;
531 
532 		if (strncmp(mntent->mnt_type, "bpf", 3) != 0)
533 			continue;
534 		err = nftw(path, do_build_table_cb, nopenfd, flags);
535 		if (err)
536 			break;
537 	}
538 	fclose(mntfile);
539 	return err;
540 }
541 
542 void delete_pinned_obj_table(struct hashmap *map)
543 {
544 	struct hashmap_entry *entry;
545 	size_t bkt;
546 
547 	if (!map)
548 		return;
549 
550 	hashmap__for_each_entry(map, entry, bkt)
551 		free(entry->pvalue);
552 
553 	hashmap__free(map);
554 }
555 
556 unsigned int get_page_size(void)
557 {
558 	static int result;
559 
560 	if (!result)
561 		result = getpagesize();
562 	return result;
563 }
564 
565 unsigned int get_possible_cpus(void)
566 {
567 	int cpus = libbpf_num_possible_cpus();
568 
569 	if (cpus < 0) {
570 		p_err("Can't get # of possible cpus: %s", strerror(-cpus));
571 		exit(-1);
572 	}
573 	return cpus;
574 }
575 
576 static char *
577 ifindex_to_name_ns(__u32 ifindex, __u32 ns_dev, __u32 ns_ino, char *buf)
578 {
579 	struct stat st;
580 	int err;
581 
582 	err = stat("/proc/self/ns/net", &st);
583 	if (err) {
584 		p_err("Can't stat /proc/self: %s", strerror(errno));
585 		return NULL;
586 	}
587 
588 	if (st.st_dev != ns_dev || st.st_ino != ns_ino)
589 		return NULL;
590 
591 	return if_indextoname(ifindex, buf);
592 }
593 
594 static int read_sysfs_hex_int(char *path)
595 {
596 	char vendor_id_buf[8];
597 	int len;
598 	int fd;
599 
600 	fd = open(path, O_RDONLY);
601 	if (fd < 0) {
602 		p_err("Can't open %s: %s", path, strerror(errno));
603 		return -1;
604 	}
605 
606 	len = read(fd, vendor_id_buf, sizeof(vendor_id_buf));
607 	close(fd);
608 	if (len < 0) {
609 		p_err("Can't read %s: %s", path, strerror(errno));
610 		return -1;
611 	}
612 	if (len >= (int)sizeof(vendor_id_buf)) {
613 		p_err("Value in %s too long", path);
614 		return -1;
615 	}
616 
617 	vendor_id_buf[len] = 0;
618 
619 	return strtol(vendor_id_buf, NULL, 0);
620 }
621 
622 static int read_sysfs_netdev_hex_int(char *devname, const char *entry_name)
623 {
624 	char full_path[64];
625 
626 	snprintf(full_path, sizeof(full_path), "/sys/class/net/%s/device/%s",
627 		 devname, entry_name);
628 
629 	return read_sysfs_hex_int(full_path);
630 }
631 
632 const char *
633 ifindex_to_arch(__u32 ifindex, __u64 ns_dev, __u64 ns_ino, const char **opt)
634 {
635 	__maybe_unused int device_id;
636 	char devname[IF_NAMESIZE];
637 	int vendor_id;
638 
639 	if (!ifindex_to_name_ns(ifindex, ns_dev, ns_ino, devname)) {
640 		p_err("Can't get net device name for ifindex %d: %s", ifindex,
641 		      strerror(errno));
642 		return NULL;
643 	}
644 
645 	vendor_id = read_sysfs_netdev_hex_int(devname, "vendor");
646 	if (vendor_id < 0) {
647 		p_err("Can't get device vendor id for %s", devname);
648 		return NULL;
649 	}
650 
651 	switch (vendor_id) {
652 #ifdef HAVE_LIBBFD_SUPPORT
653 	case 0x19ee:
654 		device_id = read_sysfs_netdev_hex_int(devname, "device");
655 		if (device_id != 0x4000 &&
656 		    device_id != 0x6000 &&
657 		    device_id != 0x6003)
658 			p_info("Unknown NFP device ID, assuming it is NFP-6xxx arch");
659 		*opt = "ctx4";
660 		return "NFP-6xxx";
661 #endif /* HAVE_LIBBFD_SUPPORT */
662 	/* No NFP support in LLVM, we have no valid triple to return. */
663 	default:
664 		p_err("Can't get arch name for device vendor id 0x%04x",
665 		      vendor_id);
666 		return NULL;
667 	}
668 }
669 
670 void print_dev_plain(__u32 ifindex, __u64 ns_dev, __u64 ns_inode)
671 {
672 	char name[IF_NAMESIZE];
673 
674 	if (!ifindex)
675 		return;
676 
677 	printf("  offloaded_to ");
678 	if (ifindex_to_name_ns(ifindex, ns_dev, ns_inode, name))
679 		printf("%s", name);
680 	else
681 		printf("ifindex %u ns_dev %llu ns_ino %llu",
682 		       ifindex, ns_dev, ns_inode);
683 }
684 
685 void print_dev_json(__u32 ifindex, __u64 ns_dev, __u64 ns_inode)
686 {
687 	char name[IF_NAMESIZE];
688 
689 	if (!ifindex)
690 		return;
691 
692 	jsonw_name(json_wtr, "dev");
693 	jsonw_start_object(json_wtr);
694 	jsonw_uint_field(json_wtr, "ifindex", ifindex);
695 	jsonw_uint_field(json_wtr, "ns_dev", ns_dev);
696 	jsonw_uint_field(json_wtr, "ns_inode", ns_inode);
697 	if (ifindex_to_name_ns(ifindex, ns_dev, ns_inode, name))
698 		jsonw_string_field(json_wtr, "ifname", name);
699 	jsonw_end_object(json_wtr);
700 }
701 
702 int parse_u32_arg(int *argc, char ***argv, __u32 *val, const char *what)
703 {
704 	char *endptr;
705 
706 	NEXT_ARGP();
707 
708 	if (*val) {
709 		p_err("%s already specified", what);
710 		return -1;
711 	}
712 
713 	*val = strtoul(**argv, &endptr, 0);
714 	if (*endptr) {
715 		p_err("can't parse %s as %s", **argv, what);
716 		return -1;
717 	}
718 	NEXT_ARGP();
719 
720 	return 0;
721 }
722 
723 int __printf(2, 0)
724 print_all_levels(__maybe_unused enum libbpf_print_level level,
725 		 const char *format, va_list args)
726 {
727 	return vfprintf(stderr, format, args);
728 }
729 
730 static int prog_fd_by_nametag(void *nametag, int **fds, bool tag)
731 {
732 	char prog_name[MAX_PROG_FULL_NAME];
733 	unsigned int id = 0;
734 	int fd, nb_fds = 0;
735 	void *tmp;
736 	int err;
737 
738 	while (true) {
739 		struct bpf_prog_info info = {};
740 		__u32 len = sizeof(info);
741 
742 		err = bpf_prog_get_next_id(id, &id);
743 		if (err) {
744 			if (errno != ENOENT) {
745 				p_err("%s", strerror(errno));
746 				goto err_close_fds;
747 			}
748 			return nb_fds;
749 		}
750 
751 		fd = bpf_prog_get_fd_by_id(id);
752 		if (fd < 0) {
753 			p_err("can't get prog by id (%u): %s",
754 			      id, strerror(errno));
755 			goto err_close_fds;
756 		}
757 
758 		err = bpf_obj_get_info_by_fd(fd, &info, &len);
759 		if (err) {
760 			p_err("can't get prog info (%u): %s",
761 			      id, strerror(errno));
762 			goto err_close_fd;
763 		}
764 
765 		if (tag && memcmp(nametag, info.tag, BPF_TAG_SIZE)) {
766 			close(fd);
767 			continue;
768 		}
769 
770 		if (!tag) {
771 			get_prog_full_name(&info, fd, prog_name,
772 					   sizeof(prog_name));
773 			if (strncmp(nametag, prog_name, sizeof(prog_name))) {
774 				close(fd);
775 				continue;
776 			}
777 		}
778 
779 		if (nb_fds > 0) {
780 			tmp = realloc(*fds, (nb_fds + 1) * sizeof(int));
781 			if (!tmp) {
782 				p_err("failed to realloc");
783 				goto err_close_fd;
784 			}
785 			*fds = tmp;
786 		}
787 		(*fds)[nb_fds++] = fd;
788 	}
789 
790 err_close_fd:
791 	close(fd);
792 err_close_fds:
793 	while (--nb_fds >= 0)
794 		close((*fds)[nb_fds]);
795 	return -1;
796 }
797 
798 int prog_parse_fds(int *argc, char ***argv, int **fds)
799 {
800 	if (is_prefix(**argv, "id")) {
801 		unsigned int id;
802 		char *endptr;
803 
804 		NEXT_ARGP();
805 
806 		id = strtoul(**argv, &endptr, 0);
807 		if (*endptr) {
808 			p_err("can't parse %s as ID", **argv);
809 			return -1;
810 		}
811 		NEXT_ARGP();
812 
813 		(*fds)[0] = bpf_prog_get_fd_by_id(id);
814 		if ((*fds)[0] < 0) {
815 			p_err("get by id (%u): %s", id, strerror(errno));
816 			return -1;
817 		}
818 		return 1;
819 	} else if (is_prefix(**argv, "tag")) {
820 		unsigned char tag[BPF_TAG_SIZE];
821 
822 		NEXT_ARGP();
823 
824 		if (sscanf(**argv, BPF_TAG_FMT, tag, tag + 1, tag + 2,
825 			   tag + 3, tag + 4, tag + 5, tag + 6, tag + 7)
826 		    != BPF_TAG_SIZE) {
827 			p_err("can't parse tag");
828 			return -1;
829 		}
830 		NEXT_ARGP();
831 
832 		return prog_fd_by_nametag(tag, fds, true);
833 	} else if (is_prefix(**argv, "name")) {
834 		char *name;
835 
836 		NEXT_ARGP();
837 
838 		name = **argv;
839 		if (strlen(name) > MAX_PROG_FULL_NAME - 1) {
840 			p_err("can't parse name");
841 			return -1;
842 		}
843 		NEXT_ARGP();
844 
845 		return prog_fd_by_nametag(name, fds, false);
846 	} else if (is_prefix(**argv, "pinned")) {
847 		char *path;
848 
849 		NEXT_ARGP();
850 
851 		path = **argv;
852 		NEXT_ARGP();
853 
854 		(*fds)[0] = open_obj_pinned_any(path, BPF_OBJ_PROG);
855 		if ((*fds)[0] < 0)
856 			return -1;
857 		return 1;
858 	}
859 
860 	p_err("expected 'id', 'tag', 'name' or 'pinned', got: '%s'?", **argv);
861 	return -1;
862 }
863 
864 int prog_parse_fd(int *argc, char ***argv)
865 {
866 	int *fds = NULL;
867 	int nb_fds, fd;
868 
869 	fds = malloc(sizeof(int));
870 	if (!fds) {
871 		p_err("mem alloc failed");
872 		return -1;
873 	}
874 	nb_fds = prog_parse_fds(argc, argv, &fds);
875 	if (nb_fds != 1) {
876 		if (nb_fds > 1) {
877 			p_err("several programs match this handle");
878 			while (nb_fds--)
879 				close(fds[nb_fds]);
880 		}
881 		fd = -1;
882 		goto exit_free;
883 	}
884 
885 	fd = fds[0];
886 exit_free:
887 	free(fds);
888 	return fd;
889 }
890 
891 static int map_fd_by_name(char *name, int **fds)
892 {
893 	unsigned int id = 0;
894 	int fd, nb_fds = 0;
895 	void *tmp;
896 	int err;
897 
898 	while (true) {
899 		struct bpf_map_info info = {};
900 		__u32 len = sizeof(info);
901 
902 		err = bpf_map_get_next_id(id, &id);
903 		if (err) {
904 			if (errno != ENOENT) {
905 				p_err("%s", strerror(errno));
906 				goto err_close_fds;
907 			}
908 			return nb_fds;
909 		}
910 
911 		fd = bpf_map_get_fd_by_id(id);
912 		if (fd < 0) {
913 			p_err("can't get map by id (%u): %s",
914 			      id, strerror(errno));
915 			goto err_close_fds;
916 		}
917 
918 		err = bpf_obj_get_info_by_fd(fd, &info, &len);
919 		if (err) {
920 			p_err("can't get map info (%u): %s",
921 			      id, strerror(errno));
922 			goto err_close_fd;
923 		}
924 
925 		if (strncmp(name, info.name, BPF_OBJ_NAME_LEN)) {
926 			close(fd);
927 			continue;
928 		}
929 
930 		if (nb_fds > 0) {
931 			tmp = realloc(*fds, (nb_fds + 1) * sizeof(int));
932 			if (!tmp) {
933 				p_err("failed to realloc");
934 				goto err_close_fd;
935 			}
936 			*fds = tmp;
937 		}
938 		(*fds)[nb_fds++] = fd;
939 	}
940 
941 err_close_fd:
942 	close(fd);
943 err_close_fds:
944 	while (--nb_fds >= 0)
945 		close((*fds)[nb_fds]);
946 	return -1;
947 }
948 
949 int map_parse_fds(int *argc, char ***argv, int **fds)
950 {
951 	if (is_prefix(**argv, "id")) {
952 		unsigned int id;
953 		char *endptr;
954 
955 		NEXT_ARGP();
956 
957 		id = strtoul(**argv, &endptr, 0);
958 		if (*endptr) {
959 			p_err("can't parse %s as ID", **argv);
960 			return -1;
961 		}
962 		NEXT_ARGP();
963 
964 		(*fds)[0] = bpf_map_get_fd_by_id(id);
965 		if ((*fds)[0] < 0) {
966 			p_err("get map by id (%u): %s", id, strerror(errno));
967 			return -1;
968 		}
969 		return 1;
970 	} else if (is_prefix(**argv, "name")) {
971 		char *name;
972 
973 		NEXT_ARGP();
974 
975 		name = **argv;
976 		if (strlen(name) > BPF_OBJ_NAME_LEN - 1) {
977 			p_err("can't parse name");
978 			return -1;
979 		}
980 		NEXT_ARGP();
981 
982 		return map_fd_by_name(name, fds);
983 	} else if (is_prefix(**argv, "pinned")) {
984 		char *path;
985 
986 		NEXT_ARGP();
987 
988 		path = **argv;
989 		NEXT_ARGP();
990 
991 		(*fds)[0] = open_obj_pinned_any(path, BPF_OBJ_MAP);
992 		if ((*fds)[0] < 0)
993 			return -1;
994 		return 1;
995 	}
996 
997 	p_err("expected 'id', 'name' or 'pinned', got: '%s'?", **argv);
998 	return -1;
999 }
1000 
1001 int map_parse_fd(int *argc, char ***argv)
1002 {
1003 	int *fds = NULL;
1004 	int nb_fds, fd;
1005 
1006 	fds = malloc(sizeof(int));
1007 	if (!fds) {
1008 		p_err("mem alloc failed");
1009 		return -1;
1010 	}
1011 	nb_fds = map_parse_fds(argc, argv, &fds);
1012 	if (nb_fds != 1) {
1013 		if (nb_fds > 1) {
1014 			p_err("several maps match this handle");
1015 			while (nb_fds--)
1016 				close(fds[nb_fds]);
1017 		}
1018 		fd = -1;
1019 		goto exit_free;
1020 	}
1021 
1022 	fd = fds[0];
1023 exit_free:
1024 	free(fds);
1025 	return fd;
1026 }
1027 
1028 int map_parse_fd_and_info(int *argc, char ***argv, void *info, __u32 *info_len)
1029 {
1030 	int err;
1031 	int fd;
1032 
1033 	fd = map_parse_fd(argc, argv);
1034 	if (fd < 0)
1035 		return -1;
1036 
1037 	err = bpf_obj_get_info_by_fd(fd, info, info_len);
1038 	if (err) {
1039 		p_err("can't get map info: %s", strerror(errno));
1040 		close(fd);
1041 		return err;
1042 	}
1043 
1044 	return fd;
1045 }
1046 
1047 size_t hash_fn_for_key_as_id(long key, void *ctx)
1048 {
1049 	return key;
1050 }
1051 
1052 bool equal_fn_for_key_as_id(long k1, long k2, void *ctx)
1053 {
1054 	return k1 == k2;
1055 }
1056 
1057 const char *bpf_attach_type_input_str(enum bpf_attach_type t)
1058 {
1059 	switch (t) {
1060 	case BPF_CGROUP_INET_INGRESS:		return "ingress";
1061 	case BPF_CGROUP_INET_EGRESS:		return "egress";
1062 	case BPF_CGROUP_INET_SOCK_CREATE:	return "sock_create";
1063 	case BPF_CGROUP_INET_SOCK_RELEASE:	return "sock_release";
1064 	case BPF_CGROUP_SOCK_OPS:		return "sock_ops";
1065 	case BPF_CGROUP_DEVICE:			return "device";
1066 	case BPF_CGROUP_INET4_BIND:		return "bind4";
1067 	case BPF_CGROUP_INET6_BIND:		return "bind6";
1068 	case BPF_CGROUP_INET4_CONNECT:		return "connect4";
1069 	case BPF_CGROUP_INET6_CONNECT:		return "connect6";
1070 	case BPF_CGROUP_INET4_POST_BIND:	return "post_bind4";
1071 	case BPF_CGROUP_INET6_POST_BIND:	return "post_bind6";
1072 	case BPF_CGROUP_INET4_GETPEERNAME:	return "getpeername4";
1073 	case BPF_CGROUP_INET6_GETPEERNAME:	return "getpeername6";
1074 	case BPF_CGROUP_INET4_GETSOCKNAME:	return "getsockname4";
1075 	case BPF_CGROUP_INET6_GETSOCKNAME:	return "getsockname6";
1076 	case BPF_CGROUP_UDP4_SENDMSG:		return "sendmsg4";
1077 	case BPF_CGROUP_UDP6_SENDMSG:		return "sendmsg6";
1078 	case BPF_CGROUP_SYSCTL:			return "sysctl";
1079 	case BPF_CGROUP_UDP4_RECVMSG:		return "recvmsg4";
1080 	case BPF_CGROUP_UDP6_RECVMSG:		return "recvmsg6";
1081 	case BPF_CGROUP_GETSOCKOPT:		return "getsockopt";
1082 	case BPF_CGROUP_SETSOCKOPT:		return "setsockopt";
1083 	case BPF_TRACE_RAW_TP:			return "raw_tp";
1084 	case BPF_TRACE_FENTRY:			return "fentry";
1085 	case BPF_TRACE_FEXIT:			return "fexit";
1086 	case BPF_MODIFY_RETURN:			return "mod_ret";
1087 	case BPF_SK_REUSEPORT_SELECT:		return "sk_skb_reuseport_select";
1088 	case BPF_SK_REUSEPORT_SELECT_OR_MIGRATE:	return "sk_skb_reuseport_select_or_migrate";
1089 	default:	return libbpf_bpf_attach_type_str(t);
1090 	}
1091 }
1092