1bd4aed0eSJiong Wang #include <stddef.h>
2bd4aed0eSJiong Wang #include <inttypes.h>
3bd4aed0eSJiong Wang #include <errno.h>
4bd4aed0eSJiong Wang #include <linux/seg6_local.h>
5bd4aed0eSJiong Wang #include <linux/bpf.h>
63e689141SToke Høiland-Jørgensen #include <bpf/bpf_helpers.h>
73e689141SToke Høiland-Jørgensen #include <bpf/bpf_endian.h>
8bd4aed0eSJiong Wang 
9bd4aed0eSJiong Wang /* Packet parsing state machine helpers. */
10bd4aed0eSJiong Wang #define cursor_advance(_cursor, _len) \
11bd4aed0eSJiong Wang 	({ void *_tmp = _cursor; _cursor += _len; _tmp; })
12bd4aed0eSJiong Wang 
13bd4aed0eSJiong Wang #define SR6_FLAG_ALERT (1 << 4)
14bd4aed0eSJiong Wang 
15bd4aed0eSJiong Wang #define BPF_PACKET_HEADER __attribute__((packed))
16bd4aed0eSJiong Wang 
17bd4aed0eSJiong Wang struct ip6_t {
18bd4aed0eSJiong Wang 	unsigned int ver:4;
19bd4aed0eSJiong Wang 	unsigned int priority:8;
20bd4aed0eSJiong Wang 	unsigned int flow_label:20;
21bd4aed0eSJiong Wang 	unsigned short payload_len;
22bd4aed0eSJiong Wang 	unsigned char next_header;
23bd4aed0eSJiong Wang 	unsigned char hop_limit;
24bd4aed0eSJiong Wang 	unsigned long long src_hi;
25bd4aed0eSJiong Wang 	unsigned long long src_lo;
26bd4aed0eSJiong Wang 	unsigned long long dst_hi;
27bd4aed0eSJiong Wang 	unsigned long long dst_lo;
28bd4aed0eSJiong Wang } BPF_PACKET_HEADER;
29bd4aed0eSJiong Wang 
30bd4aed0eSJiong Wang struct ip6_addr_t {
31bd4aed0eSJiong Wang 	unsigned long long hi;
32bd4aed0eSJiong Wang 	unsigned long long lo;
33bd4aed0eSJiong Wang } BPF_PACKET_HEADER;
34bd4aed0eSJiong Wang 
35bd4aed0eSJiong Wang struct ip6_srh_t {
36bd4aed0eSJiong Wang 	unsigned char nexthdr;
37bd4aed0eSJiong Wang 	unsigned char hdrlen;
38bd4aed0eSJiong Wang 	unsigned char type;
39bd4aed0eSJiong Wang 	unsigned char segments_left;
40bd4aed0eSJiong Wang 	unsigned char first_segment;
41bd4aed0eSJiong Wang 	unsigned char flags;
42bd4aed0eSJiong Wang 	unsigned short tag;
43bd4aed0eSJiong Wang 
44bd4aed0eSJiong Wang 	struct ip6_addr_t segments[0];
45bd4aed0eSJiong Wang } BPF_PACKET_HEADER;
46bd4aed0eSJiong Wang 
47bd4aed0eSJiong Wang struct sr6_tlv_t {
48bd4aed0eSJiong Wang 	unsigned char type;
49bd4aed0eSJiong Wang 	unsigned char len;
50bd4aed0eSJiong Wang 	unsigned char value[0];
51bd4aed0eSJiong Wang } BPF_PACKET_HEADER;
52bd4aed0eSJiong Wang 
get_srh(struct __sk_buff * skb)5311aca65eSJiri Benc static __always_inline struct ip6_srh_t *get_srh(struct __sk_buff *skb)
54bd4aed0eSJiong Wang {
55bd4aed0eSJiong Wang 	void *cursor, *data_end;
56bd4aed0eSJiong Wang 	struct ip6_srh_t *srh;
57bd4aed0eSJiong Wang 	struct ip6_t *ip;
58bd4aed0eSJiong Wang 	uint8_t *ipver;
59bd4aed0eSJiong Wang 
60bd4aed0eSJiong Wang 	data_end = (void *)(long)skb->data_end;
61bd4aed0eSJiong Wang 	cursor = (void *)(long)skb->data;
62bd4aed0eSJiong Wang 	ipver = (uint8_t *)cursor;
63bd4aed0eSJiong Wang 
64bd4aed0eSJiong Wang 	if ((void *)ipver + sizeof(*ipver) > data_end)
65bd4aed0eSJiong Wang 		return NULL;
66bd4aed0eSJiong Wang 
67bd4aed0eSJiong Wang 	if ((*ipver >> 4) != 6)
68bd4aed0eSJiong Wang 		return NULL;
69bd4aed0eSJiong Wang 
70bd4aed0eSJiong Wang 	ip = cursor_advance(cursor, sizeof(*ip));
71bd4aed0eSJiong Wang 	if ((void *)ip + sizeof(*ip) > data_end)
72bd4aed0eSJiong Wang 		return NULL;
73bd4aed0eSJiong Wang 
74bd4aed0eSJiong Wang 	if (ip->next_header != 43)
75bd4aed0eSJiong Wang 		return NULL;
76bd4aed0eSJiong Wang 
77bd4aed0eSJiong Wang 	srh = cursor_advance(cursor, sizeof(*srh));
78bd4aed0eSJiong Wang 	if ((void *)srh + sizeof(*srh) > data_end)
79bd4aed0eSJiong Wang 		return NULL;
80bd4aed0eSJiong Wang 
81bd4aed0eSJiong Wang 	if (srh->type != 4)
82bd4aed0eSJiong Wang 		return NULL;
83bd4aed0eSJiong Wang 
84bd4aed0eSJiong Wang 	return srh;
85bd4aed0eSJiong Wang }
86bd4aed0eSJiong Wang 
8711aca65eSJiri Benc static __always_inline
update_tlv_pad(struct __sk_buff * skb,uint32_t new_pad,uint32_t old_pad,uint32_t pad_off)88bd4aed0eSJiong Wang int update_tlv_pad(struct __sk_buff *skb, uint32_t new_pad,
89bd4aed0eSJiong Wang 		   uint32_t old_pad, uint32_t pad_off)
90bd4aed0eSJiong Wang {
91bd4aed0eSJiong Wang 	int err;
92bd4aed0eSJiong Wang 
93bd4aed0eSJiong Wang 	if (new_pad != old_pad) {
94bd4aed0eSJiong Wang 		err = bpf_lwt_seg6_adjust_srh(skb, pad_off,
95bd4aed0eSJiong Wang 					  (int) new_pad - (int) old_pad);
96bd4aed0eSJiong Wang 		if (err)
97bd4aed0eSJiong Wang 			return err;
98bd4aed0eSJiong Wang 	}
99bd4aed0eSJiong Wang 
100bd4aed0eSJiong Wang 	if (new_pad > 0) {
101bd4aed0eSJiong Wang 		char pad_tlv_buf[16] = {0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
102bd4aed0eSJiong Wang 					0, 0, 0};
103bd4aed0eSJiong Wang 		struct sr6_tlv_t *pad_tlv = (struct sr6_tlv_t *) pad_tlv_buf;
104bd4aed0eSJiong Wang 
105bd4aed0eSJiong Wang 		pad_tlv->type = SR6_TLV_PADDING;
106bd4aed0eSJiong Wang 		pad_tlv->len = new_pad - 2;
107bd4aed0eSJiong Wang 
108bd4aed0eSJiong Wang 		err = bpf_lwt_seg6_store_bytes(skb, pad_off,
109bd4aed0eSJiong Wang 					       (void *)pad_tlv_buf, new_pad);
110bd4aed0eSJiong Wang 		if (err)
111bd4aed0eSJiong Wang 			return err;
112bd4aed0eSJiong Wang 	}
113bd4aed0eSJiong Wang 
114bd4aed0eSJiong Wang 	return 0;
115bd4aed0eSJiong Wang }
116bd4aed0eSJiong Wang 
11711aca65eSJiri Benc static __always_inline
is_valid_tlv_boundary(struct __sk_buff * skb,struct ip6_srh_t * srh,uint32_t * tlv_off,uint32_t * pad_size,uint32_t * pad_off)118bd4aed0eSJiong Wang int is_valid_tlv_boundary(struct __sk_buff *skb, struct ip6_srh_t *srh,
119bd4aed0eSJiong Wang 			  uint32_t *tlv_off, uint32_t *pad_size,
120bd4aed0eSJiong Wang 			  uint32_t *pad_off)
121bd4aed0eSJiong Wang {
122bd4aed0eSJiong Wang 	uint32_t srh_off, cur_off;
123bd4aed0eSJiong Wang 	int offset_valid = 0;
124bd4aed0eSJiong Wang 	int err;
125bd4aed0eSJiong Wang 
126bd4aed0eSJiong Wang 	srh_off = (char *)srh - (char *)(long)skb->data;
127bd4aed0eSJiong Wang 	// cur_off = end of segments, start of possible TLVs
128bd4aed0eSJiong Wang 	cur_off = srh_off + sizeof(*srh) +
129bd4aed0eSJiong Wang 		sizeof(struct ip6_addr_t) * (srh->first_segment + 1);
130bd4aed0eSJiong Wang 
131bd4aed0eSJiong Wang 	*pad_off = 0;
132bd4aed0eSJiong Wang 
133bd4aed0eSJiong Wang 	// we can only go as far as ~10 TLVs due to the BPF max stack size
134bd4aed0eSJiong Wang 	#pragma clang loop unroll(full)
135bd4aed0eSJiong Wang 	for (int i = 0; i < 10; i++) {
136bd4aed0eSJiong Wang 		struct sr6_tlv_t tlv;
137bd4aed0eSJiong Wang 
138bd4aed0eSJiong Wang 		if (cur_off == *tlv_off)
139bd4aed0eSJiong Wang 			offset_valid = 1;
140bd4aed0eSJiong Wang 
141bd4aed0eSJiong Wang 		if (cur_off >= srh_off + ((srh->hdrlen + 1) << 3))
142bd4aed0eSJiong Wang 			break;
143bd4aed0eSJiong Wang 
144bd4aed0eSJiong Wang 		err = bpf_skb_load_bytes(skb, cur_off, &tlv, sizeof(tlv));
145bd4aed0eSJiong Wang 		if (err)
146bd4aed0eSJiong Wang 			return err;
147bd4aed0eSJiong Wang 
148bd4aed0eSJiong Wang 		if (tlv.type == SR6_TLV_PADDING) {
149bd4aed0eSJiong Wang 			*pad_size = tlv.len + sizeof(tlv);
150bd4aed0eSJiong Wang 			*pad_off = cur_off;
151bd4aed0eSJiong Wang 
152bd4aed0eSJiong Wang 			if (*tlv_off == srh_off) {
153bd4aed0eSJiong Wang 				*tlv_off = cur_off;
154bd4aed0eSJiong Wang 				offset_valid = 1;
155bd4aed0eSJiong Wang 			}
156bd4aed0eSJiong Wang 			break;
157bd4aed0eSJiong Wang 
158bd4aed0eSJiong Wang 		} else if (tlv.type == SR6_TLV_HMAC) {
159bd4aed0eSJiong Wang 			break;
160bd4aed0eSJiong Wang 		}
161bd4aed0eSJiong Wang 
162bd4aed0eSJiong Wang 		cur_off += sizeof(tlv) + tlv.len;
163bd4aed0eSJiong Wang 	} // we reached the padding or HMAC TLVs, or the end of the SRH
164bd4aed0eSJiong Wang 
165bd4aed0eSJiong Wang 	if (*pad_off == 0)
166bd4aed0eSJiong Wang 		*pad_off = cur_off;
167bd4aed0eSJiong Wang 
168bd4aed0eSJiong Wang 	if (*tlv_off == -1)
169bd4aed0eSJiong Wang 		*tlv_off = cur_off;
170bd4aed0eSJiong Wang 	else if (!offset_valid)
171bd4aed0eSJiong Wang 		return -EINVAL;
172bd4aed0eSJiong Wang 
173bd4aed0eSJiong Wang 	return 0;
174bd4aed0eSJiong Wang }
175bd4aed0eSJiong Wang 
17611aca65eSJiri Benc static __always_inline
add_tlv(struct __sk_buff * skb,struct ip6_srh_t * srh,uint32_t tlv_off,struct sr6_tlv_t * itlv,uint8_t tlv_size)177bd4aed0eSJiong Wang int add_tlv(struct __sk_buff *skb, struct ip6_srh_t *srh, uint32_t tlv_off,
178bd4aed0eSJiong Wang 	    struct sr6_tlv_t *itlv, uint8_t tlv_size)
179bd4aed0eSJiong Wang {
180bd4aed0eSJiong Wang 	uint32_t srh_off = (char *)srh - (char *)(long)skb->data;
181bd4aed0eSJiong Wang 	uint8_t len_remaining, new_pad;
182bd4aed0eSJiong Wang 	uint32_t pad_off = 0;
183bd4aed0eSJiong Wang 	uint32_t pad_size = 0;
184bd4aed0eSJiong Wang 	uint32_t partial_srh_len;
185bd4aed0eSJiong Wang 	int err;
186bd4aed0eSJiong Wang 
187bd4aed0eSJiong Wang 	if (tlv_off != -1)
188bd4aed0eSJiong Wang 		tlv_off += srh_off;
189bd4aed0eSJiong Wang 
190bd4aed0eSJiong Wang 	if (itlv->type == SR6_TLV_PADDING || itlv->type == SR6_TLV_HMAC)
191bd4aed0eSJiong Wang 		return -EINVAL;
192bd4aed0eSJiong Wang 
193bd4aed0eSJiong Wang 	err = is_valid_tlv_boundary(skb, srh, &tlv_off, &pad_size, &pad_off);
194bd4aed0eSJiong Wang 	if (err)
195bd4aed0eSJiong Wang 		return err;
196bd4aed0eSJiong Wang 
197bd4aed0eSJiong Wang 	err = bpf_lwt_seg6_adjust_srh(skb, tlv_off, sizeof(*itlv) + itlv->len);
198bd4aed0eSJiong Wang 	if (err)
199bd4aed0eSJiong Wang 		return err;
200bd4aed0eSJiong Wang 
201bd4aed0eSJiong Wang 	err = bpf_lwt_seg6_store_bytes(skb, tlv_off, (void *)itlv, tlv_size);
202bd4aed0eSJiong Wang 	if (err)
203bd4aed0eSJiong Wang 		return err;
204bd4aed0eSJiong Wang 
205bd4aed0eSJiong Wang 	// the following can't be moved inside update_tlv_pad because the
206bd4aed0eSJiong Wang 	// bpf verifier has some issues with it
207bd4aed0eSJiong Wang 	pad_off += sizeof(*itlv) + itlv->len;
208bd4aed0eSJiong Wang 	partial_srh_len = pad_off - srh_off;
209bd4aed0eSJiong Wang 	len_remaining = partial_srh_len % 8;
210bd4aed0eSJiong Wang 	new_pad = 8 - len_remaining;
211bd4aed0eSJiong Wang 
212bd4aed0eSJiong Wang 	if (new_pad == 1) // cannot pad for 1 byte only
213bd4aed0eSJiong Wang 		new_pad = 9;
214bd4aed0eSJiong Wang 	else if (new_pad == 8)
215bd4aed0eSJiong Wang 		new_pad = 0;
216bd4aed0eSJiong Wang 
217bd4aed0eSJiong Wang 	return update_tlv_pad(skb, new_pad, pad_size, pad_off);
218bd4aed0eSJiong Wang }
219bd4aed0eSJiong Wang 
22011aca65eSJiri Benc static __always_inline
delete_tlv(struct __sk_buff * skb,struct ip6_srh_t * srh,uint32_t tlv_off)221bd4aed0eSJiong Wang int delete_tlv(struct __sk_buff *skb, struct ip6_srh_t *srh,
222bd4aed0eSJiong Wang 	       uint32_t tlv_off)
223bd4aed0eSJiong Wang {
224bd4aed0eSJiong Wang 	uint32_t srh_off = (char *)srh - (char *)(long)skb->data;
225bd4aed0eSJiong Wang 	uint8_t len_remaining, new_pad;
226bd4aed0eSJiong Wang 	uint32_t partial_srh_len;
227bd4aed0eSJiong Wang 	uint32_t pad_off = 0;
228bd4aed0eSJiong Wang 	uint32_t pad_size = 0;
229bd4aed0eSJiong Wang 	struct sr6_tlv_t tlv;
230bd4aed0eSJiong Wang 	int err;
231bd4aed0eSJiong Wang 
232bd4aed0eSJiong Wang 	tlv_off += srh_off;
233bd4aed0eSJiong Wang 
234bd4aed0eSJiong Wang 	err = is_valid_tlv_boundary(skb, srh, &tlv_off, &pad_size, &pad_off);
235bd4aed0eSJiong Wang 	if (err)
236bd4aed0eSJiong Wang 		return err;
237bd4aed0eSJiong Wang 
238bd4aed0eSJiong Wang 	err = bpf_skb_load_bytes(skb, tlv_off, &tlv, sizeof(tlv));
239bd4aed0eSJiong Wang 	if (err)
240bd4aed0eSJiong Wang 		return err;
241bd4aed0eSJiong Wang 
242bd4aed0eSJiong Wang 	err = bpf_lwt_seg6_adjust_srh(skb, tlv_off, -(sizeof(tlv) + tlv.len));
243bd4aed0eSJiong Wang 	if (err)
244bd4aed0eSJiong Wang 		return err;
245bd4aed0eSJiong Wang 
246bd4aed0eSJiong Wang 	pad_off -= sizeof(tlv) + tlv.len;
247bd4aed0eSJiong Wang 	partial_srh_len = pad_off - srh_off;
248bd4aed0eSJiong Wang 	len_remaining = partial_srh_len % 8;
249bd4aed0eSJiong Wang 	new_pad = 8 - len_remaining;
250bd4aed0eSJiong Wang 	if (new_pad == 1) // cannot pad for 1 byte only
251bd4aed0eSJiong Wang 		new_pad = 9;
252bd4aed0eSJiong Wang 	else if (new_pad == 8)
253bd4aed0eSJiong Wang 		new_pad = 0;
254bd4aed0eSJiong Wang 
255bd4aed0eSJiong Wang 	return update_tlv_pad(skb, new_pad, pad_size, pad_off);
256bd4aed0eSJiong Wang }
257bd4aed0eSJiong Wang 
25811aca65eSJiri Benc static __always_inline
has_egr_tlv(struct __sk_buff * skb,struct ip6_srh_t * srh)259bd4aed0eSJiong Wang int has_egr_tlv(struct __sk_buff *skb, struct ip6_srh_t *srh)
260bd4aed0eSJiong Wang {
261bd4aed0eSJiong Wang 	int tlv_offset = sizeof(struct ip6_t) + sizeof(struct ip6_srh_t) +
262bd4aed0eSJiong Wang 		((srh->first_segment + 1) << 4);
263bd4aed0eSJiong Wang 	struct sr6_tlv_t tlv;
264bd4aed0eSJiong Wang 
265bd4aed0eSJiong Wang 	if (bpf_skb_load_bytes(skb, tlv_offset, &tlv, sizeof(struct sr6_tlv_t)))
266bd4aed0eSJiong Wang 		return 0;
267bd4aed0eSJiong Wang 
268bd4aed0eSJiong Wang 	if (tlv.type == SR6_TLV_EGRESS && tlv.len == 18) {
269bd4aed0eSJiong Wang 		struct ip6_addr_t egr_addr;
270bd4aed0eSJiong Wang 
271bd4aed0eSJiong Wang 		if (bpf_skb_load_bytes(skb, tlv_offset + 4, &egr_addr, 16))
272bd4aed0eSJiong Wang 			return 0;
273bd4aed0eSJiong Wang 
274bd4aed0eSJiong Wang 		// check if egress TLV value is correct
275b9632679SIlya Leoshkevich 		if (bpf_be64_to_cpu(egr_addr.hi) == 0xfd00000000000000 &&
276b9632679SIlya Leoshkevich 		    bpf_be64_to_cpu(egr_addr.lo) == 0x4)
277bd4aed0eSJiong Wang 			return 1;
278bd4aed0eSJiong Wang 	}
279bd4aed0eSJiong Wang 
280bd4aed0eSJiong Wang 	return 0;
281bd4aed0eSJiong Wang }
282bd4aed0eSJiong Wang 
283bd4aed0eSJiong Wang // This function will push a SRH with segments fd00::1, fd00::2, fd00::3,
284bd4aed0eSJiong Wang // fd00::4
285bd4aed0eSJiong Wang SEC("encap_srh")
__encap_srh(struct __sk_buff * skb)286bd4aed0eSJiong Wang int __encap_srh(struct __sk_buff *skb)
287bd4aed0eSJiong Wang {
288bd4aed0eSJiong Wang 	unsigned long long hi = 0xfd00000000000000;
289bd4aed0eSJiong Wang 	struct ip6_addr_t *seg;
290bd4aed0eSJiong Wang 	struct ip6_srh_t *srh;
291bd4aed0eSJiong Wang 	char srh_buf[72]; // room for 4 segments
292bd4aed0eSJiong Wang 	int err;
293bd4aed0eSJiong Wang 
294bd4aed0eSJiong Wang 	srh = (struct ip6_srh_t *)srh_buf;
295bd4aed0eSJiong Wang 	srh->nexthdr = 0;
296bd4aed0eSJiong Wang 	srh->hdrlen = 8;
297bd4aed0eSJiong Wang 	srh->type = 4;
298bd4aed0eSJiong Wang 	srh->segments_left = 3;
299bd4aed0eSJiong Wang 	srh->first_segment = 3;
300bd4aed0eSJiong Wang 	srh->flags = 0;
301bd4aed0eSJiong Wang 	srh->tag = 0;
302bd4aed0eSJiong Wang 
303bd4aed0eSJiong Wang 	seg = (struct ip6_addr_t *)((char *)srh + sizeof(*srh));
304bd4aed0eSJiong Wang 
305bd4aed0eSJiong Wang 	#pragma clang loop unroll(full)
306bd4aed0eSJiong Wang 	for (unsigned long long lo = 0; lo < 4; lo++) {
307b9632679SIlya Leoshkevich 		seg->lo = bpf_cpu_to_be64(4 - lo);
308b9632679SIlya Leoshkevich 		seg->hi = bpf_cpu_to_be64(hi);
309bd4aed0eSJiong Wang 		seg = (struct ip6_addr_t *)((char *)seg + sizeof(*seg));
310bd4aed0eSJiong Wang 	}
311bd4aed0eSJiong Wang 
312bd4aed0eSJiong Wang 	err = bpf_lwt_push_encap(skb, 0, (void *)srh, sizeof(srh_buf));
313bd4aed0eSJiong Wang 	if (err)
314bd4aed0eSJiong Wang 		return BPF_DROP;
315bd4aed0eSJiong Wang 
316bd4aed0eSJiong Wang 	return BPF_REDIRECT;
317bd4aed0eSJiong Wang }
318bd4aed0eSJiong Wang 
319bd4aed0eSJiong Wang // Add an Egress TLV fc00::4, add the flag A,
320bd4aed0eSJiong Wang // and apply End.X action to fc42::1
321bd4aed0eSJiong Wang SEC("add_egr_x")
__add_egr_x(struct __sk_buff * skb)322bd4aed0eSJiong Wang int __add_egr_x(struct __sk_buff *skb)
323bd4aed0eSJiong Wang {
324bd4aed0eSJiong Wang 	unsigned long long hi = 0xfc42000000000000;
325bd4aed0eSJiong Wang 	unsigned long long lo = 0x1;
326bd4aed0eSJiong Wang 	struct ip6_srh_t *srh = get_srh(skb);
327bd4aed0eSJiong Wang 	uint8_t new_flags = SR6_FLAG_ALERT;
328bd4aed0eSJiong Wang 	struct ip6_addr_t addr;
329bd4aed0eSJiong Wang 	int err, offset;
330bd4aed0eSJiong Wang 
331bd4aed0eSJiong Wang 	if (srh == NULL)
332bd4aed0eSJiong Wang 		return BPF_DROP;
333bd4aed0eSJiong Wang 
334bd4aed0eSJiong Wang 	uint8_t tlv[20] = {2, 18, 0, 0, 0xfd, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0,
335bd4aed0eSJiong Wang 			   0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x4};
336bd4aed0eSJiong Wang 
337bd4aed0eSJiong Wang 	err = add_tlv(skb, srh, (srh->hdrlen+1) << 3,
338bd4aed0eSJiong Wang 		      (struct sr6_tlv_t *)&tlv, 20);
339bd4aed0eSJiong Wang 	if (err)
340bd4aed0eSJiong Wang 		return BPF_DROP;
341bd4aed0eSJiong Wang 
342bd4aed0eSJiong Wang 	offset = sizeof(struct ip6_t) + offsetof(struct ip6_srh_t, flags);
343bd4aed0eSJiong Wang 	err = bpf_lwt_seg6_store_bytes(skb, offset,
344bd4aed0eSJiong Wang 				       (void *)&new_flags, sizeof(new_flags));
345bd4aed0eSJiong Wang 	if (err)
346bd4aed0eSJiong Wang 		return BPF_DROP;
347bd4aed0eSJiong Wang 
348b9632679SIlya Leoshkevich 	addr.lo = bpf_cpu_to_be64(lo);
349b9632679SIlya Leoshkevich 	addr.hi = bpf_cpu_to_be64(hi);
350bd4aed0eSJiong Wang 	err = bpf_lwt_seg6_action(skb, SEG6_LOCAL_ACTION_END_X,
351bd4aed0eSJiong Wang 				  (void *)&addr, sizeof(addr));
352bd4aed0eSJiong Wang 	if (err)
353bd4aed0eSJiong Wang 		return BPF_DROP;
354bd4aed0eSJiong Wang 	return BPF_REDIRECT;
355bd4aed0eSJiong Wang }
356bd4aed0eSJiong Wang 
357bd4aed0eSJiong Wang // Pop the Egress TLV, reset the flags, change the tag 2442 and finally do a
358bd4aed0eSJiong Wang // simple End action
359bd4aed0eSJiong Wang SEC("pop_egr")
__pop_egr(struct __sk_buff * skb)360bd4aed0eSJiong Wang int __pop_egr(struct __sk_buff *skb)
361bd4aed0eSJiong Wang {
362bd4aed0eSJiong Wang 	struct ip6_srh_t *srh = get_srh(skb);
363bd4aed0eSJiong Wang 	uint16_t new_tag = bpf_htons(2442);
364bd4aed0eSJiong Wang 	uint8_t new_flags = 0;
365bd4aed0eSJiong Wang 	int err, offset;
366bd4aed0eSJiong Wang 
367bd4aed0eSJiong Wang 	if (srh == NULL)
368bd4aed0eSJiong Wang 		return BPF_DROP;
369bd4aed0eSJiong Wang 
370bd4aed0eSJiong Wang 	if (srh->flags != SR6_FLAG_ALERT)
371bd4aed0eSJiong Wang 		return BPF_DROP;
372bd4aed0eSJiong Wang 
373bd4aed0eSJiong Wang 	if (srh->hdrlen != 11) // 4 segments + Egress TLV + Padding TLV
374bd4aed0eSJiong Wang 		return BPF_DROP;
375bd4aed0eSJiong Wang 
376bd4aed0eSJiong Wang 	if (!has_egr_tlv(skb, srh))
377bd4aed0eSJiong Wang 		return BPF_DROP;
378bd4aed0eSJiong Wang 
379bd4aed0eSJiong Wang 	err = delete_tlv(skb, srh, 8 + (srh->first_segment + 1) * 16);
380bd4aed0eSJiong Wang 	if (err)
381bd4aed0eSJiong Wang 		return BPF_DROP;
382bd4aed0eSJiong Wang 
383bd4aed0eSJiong Wang 	offset = sizeof(struct ip6_t) + offsetof(struct ip6_srh_t, flags);
384bd4aed0eSJiong Wang 	if (bpf_lwt_seg6_store_bytes(skb, offset, (void *)&new_flags,
385bd4aed0eSJiong Wang 				     sizeof(new_flags)))
386bd4aed0eSJiong Wang 		return BPF_DROP;
387bd4aed0eSJiong Wang 
388bd4aed0eSJiong Wang 	offset = sizeof(struct ip6_t) + offsetof(struct ip6_srh_t, tag);
389bd4aed0eSJiong Wang 	if (bpf_lwt_seg6_store_bytes(skb, offset, (void *)&new_tag,
390bd4aed0eSJiong Wang 				     sizeof(new_tag)))
391bd4aed0eSJiong Wang 		return BPF_DROP;
392bd4aed0eSJiong Wang 
393bd4aed0eSJiong Wang 	return BPF_OK;
394bd4aed0eSJiong Wang }
395bd4aed0eSJiong Wang 
396bd4aed0eSJiong Wang // Inspect if the Egress TLV and flag have been removed, if the tag is correct,
397bd4aed0eSJiong Wang // then apply a End.T action to reach the last segment
398bd4aed0eSJiong Wang SEC("inspect_t")
__inspect_t(struct __sk_buff * skb)399bd4aed0eSJiong Wang int __inspect_t(struct __sk_buff *skb)
400bd4aed0eSJiong Wang {
401bd4aed0eSJiong Wang 	struct ip6_srh_t *srh = get_srh(skb);
402bd4aed0eSJiong Wang 	int table = 117;
403bd4aed0eSJiong Wang 	int err;
404bd4aed0eSJiong Wang 
405bd4aed0eSJiong Wang 	if (srh == NULL)
406bd4aed0eSJiong Wang 		return BPF_DROP;
407bd4aed0eSJiong Wang 
408bd4aed0eSJiong Wang 	if (srh->flags != 0)
409bd4aed0eSJiong Wang 		return BPF_DROP;
410bd4aed0eSJiong Wang 
411bd4aed0eSJiong Wang 	if (srh->tag != bpf_htons(2442))
412bd4aed0eSJiong Wang 		return BPF_DROP;
413bd4aed0eSJiong Wang 
414bd4aed0eSJiong Wang 	if (srh->hdrlen != 8) // 4 segments
415bd4aed0eSJiong Wang 		return BPF_DROP;
416bd4aed0eSJiong Wang 
417bd4aed0eSJiong Wang 	err = bpf_lwt_seg6_action(skb, SEG6_LOCAL_ACTION_END_T,
418bd4aed0eSJiong Wang 				  (void *)&table, sizeof(table));
419bd4aed0eSJiong Wang 
420bd4aed0eSJiong Wang 	if (err)
421bd4aed0eSJiong Wang 		return BPF_DROP;
422bd4aed0eSJiong Wang 
423bd4aed0eSJiong Wang 	return BPF_REDIRECT;
424bd4aed0eSJiong Wang }
425bd4aed0eSJiong Wang 
426bd4aed0eSJiong Wang char __license[] SEC("license") = "GPL";
427