sched.c (ee2708aedad00544d38dba6df88efeb4a330bd66) sched.c (ed1ad86b8527f8f864df3c182adbfcd12a445de6)
1// SPDX-License-Identifier: GPL-2.0
2/* Multipath TCP
3 *
4 * Copyright (c) 2022, SUSE.
5 */
6
7#define pr_fmt(fmt) "MPTCP: " fmt
8
9#include <linux/kernel.h>
10#include <linux/module.h>
11#include <linux/list.h>
12#include <linux/rculist.h>
13#include <linux/spinlock.h>
14#include "protocol.h"
15
16static DEFINE_SPINLOCK(mptcp_sched_list_lock);
17static LIST_HEAD(mptcp_sched_list);
18
1// SPDX-License-Identifier: GPL-2.0
2/* Multipath TCP
3 *
4 * Copyright (c) 2022, SUSE.
5 */
6
7#define pr_fmt(fmt) "MPTCP: " fmt
8
9#include <linux/kernel.h>
10#include <linux/module.h>
11#include <linux/list.h>
12#include <linux/rculist.h>
13#include <linux/spinlock.h>
14#include "protocol.h"
15
16static DEFINE_SPINLOCK(mptcp_sched_list_lock);
17static LIST_HEAD(mptcp_sched_list);
18
19static int mptcp_sched_default_get_subflow(struct mptcp_sock *msk,
20 struct mptcp_sched_data *data)
21{
22 struct sock *ssk;
23
24 ssk = data->reinject ? mptcp_subflow_get_retrans(msk) :
25 mptcp_subflow_get_send(msk);
26 if (!ssk)
27 return -EINVAL;
28
29 mptcp_subflow_set_scheduled(mptcp_subflow_ctx(ssk), true);
30 return 0;
31}
32
33static struct mptcp_sched_ops mptcp_sched_default = {
34 .get_subflow = mptcp_sched_default_get_subflow,
35 .name = "default",
36 .owner = THIS_MODULE,
37};
38
19/* Must be called with rcu read lock held */
20struct mptcp_sched_ops *mptcp_sched_find(const char *name)
21{
22 struct mptcp_sched_ops *sched, *ret = NULL;
23
24 list_for_each_entry_rcu(sched, &mptcp_sched_list, list) {
25 if (!strcmp(sched->name, name)) {
26 ret = sched;

--- 18 unchanged lines hidden (view full) ---

45 spin_unlock(&mptcp_sched_list_lock);
46
47 pr_debug("%s registered", sched->name);
48 return 0;
49}
50
51void mptcp_unregister_scheduler(struct mptcp_sched_ops *sched)
52{
39/* Must be called with rcu read lock held */
40struct mptcp_sched_ops *mptcp_sched_find(const char *name)
41{
42 struct mptcp_sched_ops *sched, *ret = NULL;
43
44 list_for_each_entry_rcu(sched, &mptcp_sched_list, list) {
45 if (!strcmp(sched->name, name)) {
46 ret = sched;

--- 18 unchanged lines hidden (view full) ---

65 spin_unlock(&mptcp_sched_list_lock);
66
67 pr_debug("%s registered", sched->name);
68 return 0;
69}
70
71void mptcp_unregister_scheduler(struct mptcp_sched_ops *sched)
72{
73 if (sched == &mptcp_sched_default)
74 return;
75
53 spin_lock(&mptcp_sched_list_lock);
54 list_del_rcu(&sched->list);
55 spin_unlock(&mptcp_sched_list_lock);
56}
57
76 spin_lock(&mptcp_sched_list_lock);
77 list_del_rcu(&sched->list);
78 spin_unlock(&mptcp_sched_list_lock);
79}
80
81void mptcp_sched_init(void)
82{
83 mptcp_register_scheduler(&mptcp_sched_default);
84}
85
58int mptcp_init_sched(struct mptcp_sock *msk,
59 struct mptcp_sched_ops *sched)
60{
61 if (!sched)
86int mptcp_init_sched(struct mptcp_sock *msk,
87 struct mptcp_sched_ops *sched)
88{
89 if (!sched)
62 goto out;
90 sched = &mptcp_sched_default;
63
64 if (!bpf_try_module_get(sched, sched->owner))
65 return -EBUSY;
66
67 msk->sched = sched;
68 if (msk->sched->init)
69 msk->sched->init(msk);
70
71 pr_debug("sched=%s", msk->sched->name);
72
91
92 if (!bpf_try_module_get(sched, sched->owner))
93 return -EBUSY;
94
95 msk->sched = sched;
96 if (msk->sched->init)
97 msk->sched->init(msk);
98
99 pr_debug("sched=%s", msk->sched->name);
100
73out:
74 return 0;
75}
76
77void mptcp_release_sched(struct mptcp_sock *msk)
78{
79 struct mptcp_sched_ops *sched = msk->sched;
80
81 if (!sched)

--- 30 unchanged lines hidden (view full) ---

112 return -EINVAL;
113 }
114
115 mptcp_for_each_subflow(msk, subflow) {
116 if (READ_ONCE(subflow->scheduled))
117 return 0;
118 }
119
101 return 0;
102}
103
104void mptcp_release_sched(struct mptcp_sock *msk)
105{
106 struct mptcp_sched_ops *sched = msk->sched;
107
108 if (!sched)

--- 30 unchanged lines hidden (view full) ---

139 return -EINVAL;
140 }
141
142 mptcp_for_each_subflow(msk, subflow) {
143 if (READ_ONCE(subflow->scheduled))
144 return 0;
145 }
146
120 if (!msk->sched) {
121 struct sock *ssk;
122
123 ssk = mptcp_subflow_get_send(msk);
124 if (!ssk)
125 return -EINVAL;
126 mptcp_subflow_set_scheduled(mptcp_subflow_ctx(ssk), true);
127 return 0;
128 }
129
130 data.reinject = false;
147 data.reinject = false;
148 if (msk->sched == &mptcp_sched_default || !msk->sched)
149 return mptcp_sched_default_get_subflow(msk, &data);
131 return msk->sched->get_subflow(msk, &data);
132}
133
134int mptcp_sched_get_retrans(struct mptcp_sock *msk)
135{
136 struct mptcp_subflow_context *subflow;
137 struct mptcp_sched_data data;
138
139 msk_owned_by_me(msk);
140
141 /* the following check is moved out of mptcp_subflow_get_retrans */
142 if (__mptcp_check_fallback(msk))
143 return -EINVAL;
144
145 mptcp_for_each_subflow(msk, subflow) {
146 if (READ_ONCE(subflow->scheduled))
147 return 0;
148 }
149
150 return msk->sched->get_subflow(msk, &data);
151}
152
153int mptcp_sched_get_retrans(struct mptcp_sock *msk)
154{
155 struct mptcp_subflow_context *subflow;
156 struct mptcp_sched_data data;
157
158 msk_owned_by_me(msk);
159
160 /* the following check is moved out of mptcp_subflow_get_retrans */
161 if (__mptcp_check_fallback(msk))
162 return -EINVAL;
163
164 mptcp_for_each_subflow(msk, subflow) {
165 if (READ_ONCE(subflow->scheduled))
166 return 0;
167 }
168
150 if (!msk->sched) {
151 struct sock *ssk;
152
153 ssk = mptcp_subflow_get_retrans(msk);
154 if (!ssk)
155 return -EINVAL;
156 mptcp_subflow_set_scheduled(mptcp_subflow_ctx(ssk), true);
157 return 0;
158 }
159
160 data.reinject = true;
169 data.reinject = true;
170 if (msk->sched == &mptcp_sched_default || !msk->sched)
171 return mptcp_sched_default_get_subflow(msk, &data);
161 return msk->sched->get_subflow(msk, &data);
162}
172 return msk->sched->get_subflow(msk, &data);
173}